var/home/core/zuul-output/0000755000175000017500000000000015112474022014523 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015112506245015472 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004544646315112506237017717 0ustar rootrootNov 29 04:38:43 crc systemd[1]: Starting Kubernetes Kubelet... Nov 29 04:38:43 crc restorecon[4758]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 29 04:38:43 crc restorecon[4758]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 29 04:38:44 crc restorecon[4758]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 29 04:38:44 crc restorecon[4758]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Nov 29 04:38:44 crc kubenswrapper[4799]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 29 04:38:44 crc kubenswrapper[4799]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Nov 29 04:38:44 crc kubenswrapper[4799]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 29 04:38:44 crc kubenswrapper[4799]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 29 04:38:44 crc kubenswrapper[4799]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Nov 29 04:38:44 crc kubenswrapper[4799]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.488183 4799 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.491697 4799 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.491721 4799 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.491728 4799 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.491735 4799 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.491740 4799 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.491745 4799 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.491751 4799 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.491763 4799 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.491769 4799 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.491773 4799 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.491778 4799 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.491783 4799 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.491805 4799 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.491811 4799 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.491816 4799 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.491821 4799 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.491826 4799 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.491831 4799 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.491836 4799 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.491840 4799 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.491845 4799 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.491850 4799 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.491854 4799 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.491862 4799 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.491868 4799 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.491873 4799 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.491878 4799 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.491883 4799 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.491889 4799 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.491893 4799 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.491898 4799 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.491903 4799 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.491908 4799 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.491913 4799 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.491918 4799 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.491923 4799 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.491928 4799 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.491934 4799 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.491938 4799 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.491943 4799 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.491949 4799 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.491954 4799 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.491959 4799 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.491964 4799 feature_gate.go:330] unrecognized feature gate: Example Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.491969 4799 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.491974 4799 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.491980 4799 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.491986 4799 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.491990 4799 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.491997 4799 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.492003 4799 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.492008 4799 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.492013 4799 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.492018 4799 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.492022 4799 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.492027 4799 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.492032 4799 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.492036 4799 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.492041 4799 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.492045 4799 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.492051 4799 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.492057 4799 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.492066 4799 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.492073 4799 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.492079 4799 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.492085 4799 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.492090 4799 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.492095 4799 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.492100 4799 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.492105 4799 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.492111 4799 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492379 4799 flags.go:64] FLAG: --address="0.0.0.0" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492397 4799 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492406 4799 flags.go:64] FLAG: --anonymous-auth="true" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492413 4799 flags.go:64] FLAG: --application-metrics-count-limit="100" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492421 4799 flags.go:64] FLAG: --authentication-token-webhook="false" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492426 4799 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492433 4799 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492440 4799 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492446 4799 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492452 4799 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492458 4799 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492464 4799 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492469 4799 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492475 4799 flags.go:64] FLAG: --cgroup-root="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492480 4799 flags.go:64] FLAG: --cgroups-per-qos="true" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492486 4799 flags.go:64] FLAG: --client-ca-file="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492491 4799 flags.go:64] FLAG: --cloud-config="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492496 4799 flags.go:64] FLAG: --cloud-provider="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492502 4799 flags.go:64] FLAG: --cluster-dns="[]" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492511 4799 flags.go:64] FLAG: --cluster-domain="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492517 4799 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492523 4799 flags.go:64] FLAG: --config-dir="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492528 4799 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492534 4799 flags.go:64] FLAG: --container-log-max-files="5" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492542 4799 flags.go:64] FLAG: --container-log-max-size="10Mi" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492548 4799 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492553 4799 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492559 4799 flags.go:64] FLAG: --containerd-namespace="k8s.io" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492565 4799 flags.go:64] FLAG: --contention-profiling="false" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492570 4799 flags.go:64] FLAG: --cpu-cfs-quota="true" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492576 4799 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492582 4799 flags.go:64] FLAG: --cpu-manager-policy="none" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492587 4799 flags.go:64] FLAG: --cpu-manager-policy-options="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492595 4799 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492606 4799 flags.go:64] FLAG: --enable-controller-attach-detach="true" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492611 4799 flags.go:64] FLAG: --enable-debugging-handlers="true" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492617 4799 flags.go:64] FLAG: --enable-load-reader="false" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492623 4799 flags.go:64] FLAG: --enable-server="true" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492628 4799 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492634 4799 flags.go:64] FLAG: --event-burst="100" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492640 4799 flags.go:64] FLAG: --event-qps="50" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492646 4799 flags.go:64] FLAG: --event-storage-age-limit="default=0" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492652 4799 flags.go:64] FLAG: --event-storage-event-limit="default=0" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492657 4799 flags.go:64] FLAG: --eviction-hard="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492664 4799 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492669 4799 flags.go:64] FLAG: --eviction-minimum-reclaim="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492675 4799 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492681 4799 flags.go:64] FLAG: --eviction-soft="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492686 4799 flags.go:64] FLAG: --eviction-soft-grace-period="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492692 4799 flags.go:64] FLAG: --exit-on-lock-contention="false" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492697 4799 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492702 4799 flags.go:64] FLAG: --experimental-mounter-path="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492708 4799 flags.go:64] FLAG: --fail-cgroupv1="false" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492713 4799 flags.go:64] FLAG: --fail-swap-on="true" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492719 4799 flags.go:64] FLAG: --feature-gates="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492725 4799 flags.go:64] FLAG: --file-check-frequency="20s" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492732 4799 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492738 4799 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492744 4799 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492750 4799 flags.go:64] FLAG: --healthz-port="10248" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492755 4799 flags.go:64] FLAG: --help="false" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492761 4799 flags.go:64] FLAG: --hostname-override="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492767 4799 flags.go:64] FLAG: --housekeeping-interval="10s" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492773 4799 flags.go:64] FLAG: --http-check-frequency="20s" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492778 4799 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492785 4799 flags.go:64] FLAG: --image-credential-provider-config="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492809 4799 flags.go:64] FLAG: --image-gc-high-threshold="85" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492815 4799 flags.go:64] FLAG: --image-gc-low-threshold="80" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492820 4799 flags.go:64] FLAG: --image-service-endpoint="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492826 4799 flags.go:64] FLAG: --kernel-memcg-notification="false" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492832 4799 flags.go:64] FLAG: --kube-api-burst="100" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492838 4799 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492844 4799 flags.go:64] FLAG: --kube-api-qps="50" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492850 4799 flags.go:64] FLAG: --kube-reserved="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492856 4799 flags.go:64] FLAG: --kube-reserved-cgroup="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492862 4799 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492867 4799 flags.go:64] FLAG: --kubelet-cgroups="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492873 4799 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492879 4799 flags.go:64] FLAG: --lock-file="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492884 4799 flags.go:64] FLAG: --log-cadvisor-usage="false" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492890 4799 flags.go:64] FLAG: --log-flush-frequency="5s" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492896 4799 flags.go:64] FLAG: --log-json-info-buffer-size="0" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492904 4799 flags.go:64] FLAG: --log-json-split-stream="false" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492909 4799 flags.go:64] FLAG: --log-text-info-buffer-size="0" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492915 4799 flags.go:64] FLAG: --log-text-split-stream="false" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492920 4799 flags.go:64] FLAG: --logging-format="text" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492926 4799 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492934 4799 flags.go:64] FLAG: --make-iptables-util-chains="true" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492940 4799 flags.go:64] FLAG: --manifest-url="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492945 4799 flags.go:64] FLAG: --manifest-url-header="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492952 4799 flags.go:64] FLAG: --max-housekeeping-interval="15s" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492958 4799 flags.go:64] FLAG: --max-open-files="1000000" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492965 4799 flags.go:64] FLAG: --max-pods="110" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492971 4799 flags.go:64] FLAG: --maximum-dead-containers="-1" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492976 4799 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492982 4799 flags.go:64] FLAG: --memory-manager-policy="None" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492987 4799 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.492994 4799 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493000 4799 flags.go:64] FLAG: --node-ip="192.168.126.11" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493006 4799 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493018 4799 flags.go:64] FLAG: --node-status-max-images="50" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493023 4799 flags.go:64] FLAG: --node-status-update-frequency="10s" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493029 4799 flags.go:64] FLAG: --oom-score-adj="-999" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493034 4799 flags.go:64] FLAG: --pod-cidr="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493040 4799 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493048 4799 flags.go:64] FLAG: --pod-manifest-path="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493055 4799 flags.go:64] FLAG: --pod-max-pids="-1" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493060 4799 flags.go:64] FLAG: --pods-per-core="0" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493066 4799 flags.go:64] FLAG: --port="10250" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493071 4799 flags.go:64] FLAG: --protect-kernel-defaults="false" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493077 4799 flags.go:64] FLAG: --provider-id="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493082 4799 flags.go:64] FLAG: --qos-reserved="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493087 4799 flags.go:64] FLAG: --read-only-port="10255" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493093 4799 flags.go:64] FLAG: --register-node="true" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493099 4799 flags.go:64] FLAG: --register-schedulable="true" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493104 4799 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493113 4799 flags.go:64] FLAG: --registry-burst="10" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493118 4799 flags.go:64] FLAG: --registry-qps="5" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493124 4799 flags.go:64] FLAG: --reserved-cpus="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493132 4799 flags.go:64] FLAG: --reserved-memory="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493139 4799 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493144 4799 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493150 4799 flags.go:64] FLAG: --rotate-certificates="false" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493155 4799 flags.go:64] FLAG: --rotate-server-certificates="false" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493161 4799 flags.go:64] FLAG: --runonce="false" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493166 4799 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493172 4799 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493177 4799 flags.go:64] FLAG: --seccomp-default="false" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493183 4799 flags.go:64] FLAG: --serialize-image-pulls="true" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493188 4799 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493199 4799 flags.go:64] FLAG: --storage-driver-db="cadvisor" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493204 4799 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493211 4799 flags.go:64] FLAG: --storage-driver-password="root" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493216 4799 flags.go:64] FLAG: --storage-driver-secure="false" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493222 4799 flags.go:64] FLAG: --storage-driver-table="stats" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493227 4799 flags.go:64] FLAG: --storage-driver-user="root" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493233 4799 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493239 4799 flags.go:64] FLAG: --sync-frequency="1m0s" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493245 4799 flags.go:64] FLAG: --system-cgroups="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493251 4799 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493260 4799 flags.go:64] FLAG: --system-reserved-cgroup="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493265 4799 flags.go:64] FLAG: --tls-cert-file="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493271 4799 flags.go:64] FLAG: --tls-cipher-suites="[]" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493278 4799 flags.go:64] FLAG: --tls-min-version="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493284 4799 flags.go:64] FLAG: --tls-private-key-file="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493290 4799 flags.go:64] FLAG: --topology-manager-policy="none" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493295 4799 flags.go:64] FLAG: --topology-manager-policy-options="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493301 4799 flags.go:64] FLAG: --topology-manager-scope="container" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493307 4799 flags.go:64] FLAG: --v="2" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493314 4799 flags.go:64] FLAG: --version="false" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493321 4799 flags.go:64] FLAG: --vmodule="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493330 4799 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.493335 4799 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493473 4799 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493481 4799 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493488 4799 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493493 4799 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493498 4799 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493504 4799 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493509 4799 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493514 4799 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493519 4799 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493528 4799 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493532 4799 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493537 4799 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493542 4799 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493548 4799 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493552 4799 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493557 4799 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493562 4799 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493566 4799 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493572 4799 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493577 4799 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493582 4799 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493587 4799 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493592 4799 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493597 4799 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493602 4799 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493607 4799 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493613 4799 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493619 4799 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493624 4799 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493629 4799 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493636 4799 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493640 4799 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493645 4799 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493651 4799 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493657 4799 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493662 4799 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493668 4799 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493673 4799 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493678 4799 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493684 4799 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493689 4799 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493695 4799 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493700 4799 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493705 4799 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493710 4799 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493715 4799 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493720 4799 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493726 4799 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493731 4799 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493737 4799 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493742 4799 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493748 4799 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493753 4799 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493758 4799 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493763 4799 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493768 4799 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493774 4799 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493779 4799 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493784 4799 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493813 4799 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493819 4799 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493825 4799 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493833 4799 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493837 4799 feature_gate.go:330] unrecognized feature gate: Example Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493842 4799 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493847 4799 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493852 4799 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493857 4799 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493862 4799 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493866 4799 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.493871 4799 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.494023 4799 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.504493 4799 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.504549 4799 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.504710 4799 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.504734 4799 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.504743 4799 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.504752 4799 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.504762 4799 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.504772 4799 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.504780 4799 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.504851 4799 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.504863 4799 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.504874 4799 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.504882 4799 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.504890 4799 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.504898 4799 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.504905 4799 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.504914 4799 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.504922 4799 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.504933 4799 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.504942 4799 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.504950 4799 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.504958 4799 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.504965 4799 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.504973 4799 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.504981 4799 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.504992 4799 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505004 4799 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505012 4799 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505021 4799 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505029 4799 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505039 4799 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505049 4799 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505057 4799 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505066 4799 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505075 4799 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505083 4799 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505092 4799 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505101 4799 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505109 4799 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505117 4799 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505125 4799 feature_gate.go:330] unrecognized feature gate: Example Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505133 4799 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505141 4799 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505149 4799 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505156 4799 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505164 4799 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505171 4799 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505181 4799 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505190 4799 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505198 4799 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505206 4799 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505214 4799 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505222 4799 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505229 4799 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505241 4799 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505252 4799 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505261 4799 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505269 4799 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505278 4799 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505287 4799 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505295 4799 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505303 4799 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505311 4799 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505320 4799 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505328 4799 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505335 4799 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505343 4799 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505351 4799 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505359 4799 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505366 4799 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505374 4799 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505382 4799 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505390 4799 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.505404 4799 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505622 4799 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505635 4799 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505643 4799 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505653 4799 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505661 4799 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505671 4799 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505681 4799 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505692 4799 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505702 4799 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505711 4799 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505720 4799 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505728 4799 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505739 4799 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505749 4799 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505758 4799 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505767 4799 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505777 4799 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505785 4799 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505829 4799 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505840 4799 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505849 4799 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505856 4799 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505864 4799 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505872 4799 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505881 4799 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505889 4799 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505897 4799 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505904 4799 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505912 4799 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505920 4799 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505928 4799 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505935 4799 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505943 4799 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505950 4799 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505958 4799 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505966 4799 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505973 4799 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505981 4799 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505988 4799 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.505996 4799 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.506006 4799 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.506014 4799 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.506022 4799 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.506029 4799 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.506037 4799 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.506045 4799 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.506053 4799 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.506061 4799 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.506069 4799 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.506076 4799 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.506084 4799 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.506092 4799 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.506101 4799 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.506109 4799 feature_gate.go:330] unrecognized feature gate: Example Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.506117 4799 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.506125 4799 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.506133 4799 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.506141 4799 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.506148 4799 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.506157 4799 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.506165 4799 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.506173 4799 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.506180 4799 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.506188 4799 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.506196 4799 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.506203 4799 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.506211 4799 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.506219 4799 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.506226 4799 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.506236 4799 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.506246 4799 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.506259 4799 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.506524 4799 server.go:940] "Client rotation is on, will bootstrap in background" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.512045 4799 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.512823 4799 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.513846 4799 server.go:997] "Starting client certificate rotation" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.513893 4799 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.514049 4799 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-10 13:28:05.545215977 +0000 UTC Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.514195 4799 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.521218 4799 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 29 04:38:44 crc kubenswrapper[4799]: E1129 04:38:44.523403 4799 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.83:6443: connect: connection refused" logger="UnhandledError" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.524993 4799 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.534578 4799 log.go:25] "Validated CRI v1 runtime API" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.555305 4799 log.go:25] "Validated CRI v1 image API" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.557316 4799 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.560103 4799 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-11-29-04-34-31-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.560161 4799 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.584130 4799 manager.go:217] Machine: {Timestamp:2025-11-29 04:38:44.582221785 +0000 UTC m=+0.225152235 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:882c4d8e-57b7-4312-8499-40d1f85d56c9 BootID:6f4e4672-da32-440c-8c4f-6e7f4942f55e Filesystems:[{Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:86:be:e1 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:86:be:e1 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:05:95:36 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:cf:a1:cb Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:47:fe:15 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:34:11:af Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:eb:d8:7a Speed:-1 Mtu:1496} {Name:eth10 MacAddress:06:53:6f:b3:7e:7c Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:4a:4c:3e:c9:ca:5a Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.584458 4799 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.584721 4799 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.585610 4799 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.586023 4799 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.586087 4799 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.586425 4799 topology_manager.go:138] "Creating topology manager with none policy" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.586445 4799 container_manager_linux.go:303] "Creating device plugin manager" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.586760 4799 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.586842 4799 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.587094 4799 state_mem.go:36] "Initialized new in-memory state store" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.587251 4799 server.go:1245] "Using root directory" path="/var/lib/kubelet" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.588161 4799 kubelet.go:418] "Attempting to sync node with API server" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.588198 4799 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.588239 4799 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.588261 4799 kubelet.go:324] "Adding apiserver pod source" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.588281 4799 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.592039 4799 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.83:6443: connect: connection refused Nov 29 04:38:44 crc kubenswrapper[4799]: E1129 04:38:44.592254 4799 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.83:6443: connect: connection refused" logger="UnhandledError" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.592516 4799 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.593017 4799 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.83:6443: connect: connection refused Nov 29 04:38:44 crc kubenswrapper[4799]: E1129 04:38:44.593119 4799 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.83:6443: connect: connection refused" logger="UnhandledError" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.593057 4799 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.594723 4799 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.595545 4799 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.595583 4799 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.595598 4799 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.595611 4799 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.595632 4799 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.595644 4799 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.595657 4799 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.595678 4799 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.595693 4799 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.595707 4799 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.595725 4799 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.595737 4799 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.596220 4799 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.596887 4799 server.go:1280] "Started kubelet" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.597203 4799 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.597407 4799 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.598412 4799 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.83:6443: connect: connection refused Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.598459 4799 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 29 04:38:44 crc systemd[1]: Started Kubernetes Kubelet. Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.599767 4799 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.599809 4799 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.600135 4799 server.go:460] "Adding debug handlers to kubelet server" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.600272 4799 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-04 05:53:39.446523245 +0000 UTC Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.600317 4799 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 865h14m54.84620841s for next certificate rotation Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.600426 4799 volume_manager.go:287] "The desired_state_of_world populator starts" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.600461 4799 volume_manager.go:289] "Starting Kubelet Volume Manager" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.600627 4799 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Nov 29 04:38:44 crc kubenswrapper[4799]: E1129 04:38:44.600435 4799 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.600963 4799 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.83:6443: connect: connection refused Nov 29 04:38:44 crc kubenswrapper[4799]: E1129 04:38:44.601031 4799 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.83:6443: connect: connection refused" logger="UnhandledError" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.601220 4799 factory.go:55] Registering systemd factory Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.601243 4799 factory.go:221] Registration of the systemd container factory successfully Nov 29 04:38:44 crc kubenswrapper[4799]: E1129 04:38:44.601391 4799 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.83:6443: connect: connection refused" interval="200ms" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.602699 4799 factory.go:153] Registering CRI-O factory Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.602743 4799 factory.go:221] Registration of the crio container factory successfully Nov 29 04:38:44 crc kubenswrapper[4799]: E1129 04:38:44.602921 4799 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.83:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187c606132726663 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-29 04:38:44.596844131 +0000 UTC m=+0.239774561,LastTimestamp:2025-11-29 04:38:44.596844131 +0000 UTC m=+0.239774561,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.608576 4799 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.608621 4799 factory.go:103] Registering Raw factory Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.608645 4799 manager.go:1196] Started watching for new ooms in manager Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.611186 4799 manager.go:319] Starting recovery of all containers Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.619782 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.620092 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.620108 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.620142 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.620164 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.620176 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.620189 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.620236 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.620255 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.620267 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.620299 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.620322 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.620337 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.620352 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.620385 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.620404 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.620416 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.620428 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.620469 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.620490 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.620511 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.620559 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.620572 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.620590 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.620635 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.620677 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.622196 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.622236 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.622251 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.622275 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.622310 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.622324 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.622337 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.622351 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.622385 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.622401 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.622415 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.622627 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.622642 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.622660 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.622673 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.622687 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.622703 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.622717 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.622733 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.622763 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.622780 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.622836 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.622857 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.622873 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.622889 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.622904 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.622925 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.622939 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.622952 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.622987 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623034 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623051 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623067 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623083 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623100 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623117 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623132 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623147 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623162 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623176 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623193 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623209 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623224 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623241 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623254 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623270 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623283 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623299 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623339 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623372 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623388 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623403 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623418 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623432 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623455 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623468 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623482 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623496 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623513 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623528 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623542 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623560 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623573 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623588 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623602 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623617 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623632 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623646 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623660 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623673 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623686 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623702 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623717 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623731 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623745 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623758 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623772 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623801 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623822 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623836 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623852 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623865 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623881 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623898 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623915 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623929 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623945 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623960 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623974 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.623987 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624027 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624041 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624076 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624089 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624101 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624113 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624127 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624158 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624171 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624182 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624196 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624208 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624220 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624231 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624242 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624256 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624269 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624282 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624295 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624307 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624320 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624331 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624345 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624358 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624369 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624381 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624392 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624403 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624414 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624424 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624435 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624447 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624459 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624470 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624481 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624493 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624503 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624516 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624530 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624545 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624559 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624573 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624588 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624602 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624616 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624630 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624647 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624662 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624678 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624693 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624708 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624722 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624737 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624754 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624767 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624782 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624839 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624854 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624872 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624886 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624902 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624919 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624935 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624952 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624967 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624981 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.624997 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.625016 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.625032 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.625047 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.625061 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.625077 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.625091 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.625731 4799 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.625759 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.625776 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.625811 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.625827 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.625845 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.625861 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.625877 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.625891 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.625906 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.625923 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.625939 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.625951 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.625963 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.625975 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.625990 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.626002 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.626014 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.626027 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.626037 4799 reconstruct.go:97] "Volume reconstruction finished" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.626044 4799 reconciler.go:26] "Reconciler: start to sync state" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.640402 4799 manager.go:324] Recovery completed Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.652300 4799 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.653768 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.656739 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.656815 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.656830 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.657496 4799 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.657531 4799 status_manager.go:217] "Starting to sync pod status with apiserver" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.657561 4799 kubelet.go:2335] "Starting kubelet main sync loop" Nov 29 04:38:44 crc kubenswrapper[4799]: E1129 04:38:44.657609 4799 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.658571 4799 cpu_manager.go:225] "Starting CPU manager" policy="none" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.658606 4799 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.658635 4799 state_mem.go:36] "Initialized new in-memory state store" Nov 29 04:38:44 crc kubenswrapper[4799]: W1129 04:38:44.658577 4799 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.83:6443: connect: connection refused Nov 29 04:38:44 crc kubenswrapper[4799]: E1129 04:38:44.658731 4799 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.83:6443: connect: connection refused" logger="UnhandledError" Nov 29 04:38:44 crc kubenswrapper[4799]: E1129 04:38:44.701294 4799 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.712361 4799 policy_none.go:49] "None policy: Start" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.713410 4799 memory_manager.go:170] "Starting memorymanager" policy="None" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.713442 4799 state_mem.go:35] "Initializing new in-memory state store" Nov 29 04:38:44 crc kubenswrapper[4799]: E1129 04:38:44.758697 4799 kubelet.go:2359] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.774282 4799 manager.go:334] "Starting Device Plugin manager" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.774341 4799 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.774359 4799 server.go:79] "Starting device plugin registration server" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.774924 4799 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.774947 4799 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.775136 4799 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.775237 4799 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.775247 4799 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 29 04:38:44 crc kubenswrapper[4799]: E1129 04:38:44.785907 4799 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 29 04:38:44 crc kubenswrapper[4799]: E1129 04:38:44.802514 4799 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.83:6443: connect: connection refused" interval="400ms" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.875881 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.877837 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.877895 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.877909 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.877942 4799 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 29 04:38:44 crc kubenswrapper[4799]: E1129 04:38:44.878521 4799 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.83:6443: connect: connection refused" node="crc" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.959658 4799 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.959844 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.961290 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.961402 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.961436 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.961862 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.962024 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.962070 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.963109 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.963162 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.963183 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.963344 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.963411 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.963422 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.963581 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.963850 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.963960 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.964375 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.964414 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.964433 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.964589 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.964774 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.964849 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.964894 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.964927 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.964938 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.965494 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.965535 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.965553 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.965720 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.965855 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.965885 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.965895 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.966099 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.966187 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.966634 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.966668 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.966683 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.966869 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.966904 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.967570 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.967614 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.967626 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.968295 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.968331 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:38:44 crc kubenswrapper[4799]: I1129 04:38:44.968348 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.030166 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.030231 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.030287 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.030352 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.030395 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.030433 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.030465 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.030488 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.030510 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.030531 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.030550 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.030569 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.030585 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.030604 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.030622 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.078904 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.080620 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.080696 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.080715 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.080760 4799 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 29 04:38:45 crc kubenswrapper[4799]: E1129 04:38:45.081782 4799 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.83:6443: connect: connection refused" node="crc" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.131453 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.131548 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.131610 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.131647 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.131689 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.131715 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.131752 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.131784 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.131835 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.131828 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.131882 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.131861 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.131764 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.131946 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.131953 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.132022 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.132138 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.132126 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.132195 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.132277 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.132293 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.132342 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.132358 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.132399 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.132422 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.132435 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.132473 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.132496 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.132524 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.132613 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 29 04:38:45 crc kubenswrapper[4799]: E1129 04:38:45.203410 4799 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.83:6443: connect: connection refused" interval="800ms" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.299945 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.307753 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 04:38:45 crc kubenswrapper[4799]: W1129 04:38:45.329838 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-f8ad6bba8b88f025af53adfd2063a9e143103007d344eab1b7b4c532c0a08026 WatchSource:0}: Error finding container f8ad6bba8b88f025af53adfd2063a9e143103007d344eab1b7b4c532c0a08026: Status 404 returned error can't find the container with id f8ad6bba8b88f025af53adfd2063a9e143103007d344eab1b7b4c532c0a08026 Nov 29 04:38:45 crc kubenswrapper[4799]: W1129 04:38:45.330865 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-2982a6d450a9e5baf844b768801039c949232b2cca27f6bc0e3d99cccf0576f0 WatchSource:0}: Error finding container 2982a6d450a9e5baf844b768801039c949232b2cca27f6bc0e3d99cccf0576f0: Status 404 returned error can't find the container with id 2982a6d450a9e5baf844b768801039c949232b2cca27f6bc0e3d99cccf0576f0 Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.332204 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 29 04:38:45 crc kubenswrapper[4799]: W1129 04:38:45.349444 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-4e75dbd2828109703f688ee79c9b4e1070338397acb59de75002b9a6155141f9 WatchSource:0}: Error finding container 4e75dbd2828109703f688ee79c9b4e1070338397acb59de75002b9a6155141f9: Status 404 returned error can't find the container with id 4e75dbd2828109703f688ee79c9b4e1070338397acb59de75002b9a6155141f9 Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.354345 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.360371 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 29 04:38:45 crc kubenswrapper[4799]: W1129 04:38:45.373296 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-ecb364e845f7037324c8d2da20a4120ef5c1558bf9e47c20963eb4b0353b65ee WatchSource:0}: Error finding container ecb364e845f7037324c8d2da20a4120ef5c1558bf9e47c20963eb4b0353b65ee: Status 404 returned error can't find the container with id ecb364e845f7037324c8d2da20a4120ef5c1558bf9e47c20963eb4b0353b65ee Nov 29 04:38:45 crc kubenswrapper[4799]: W1129 04:38:45.379100 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-9c250a70b2c436b9976291eac10ba727c6c9a3492ea3be3062bf67a8c0f9407e WatchSource:0}: Error finding container 9c250a70b2c436b9976291eac10ba727c6c9a3492ea3be3062bf67a8c0f9407e: Status 404 returned error can't find the container with id 9c250a70b2c436b9976291eac10ba727c6c9a3492ea3be3062bf67a8c0f9407e Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.482677 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.484323 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.484371 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.484385 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.484416 4799 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 29 04:38:45 crc kubenswrapper[4799]: E1129 04:38:45.484938 4799 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.83:6443: connect: connection refused" node="crc" Nov 29 04:38:45 crc kubenswrapper[4799]: W1129 04:38:45.510934 4799 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.83:6443: connect: connection refused Nov 29 04:38:45 crc kubenswrapper[4799]: E1129 04:38:45.511000 4799 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.83:6443: connect: connection refused" logger="UnhandledError" Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.599985 4799 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.83:6443: connect: connection refused Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.661436 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f8ad6bba8b88f025af53adfd2063a9e143103007d344eab1b7b4c532c0a08026"} Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.662943 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"2982a6d450a9e5baf844b768801039c949232b2cca27f6bc0e3d99cccf0576f0"} Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.664965 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"9c250a70b2c436b9976291eac10ba727c6c9a3492ea3be3062bf67a8c0f9407e"} Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.666050 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"ecb364e845f7037324c8d2da20a4120ef5c1558bf9e47c20963eb4b0353b65ee"} Nov 29 04:38:45 crc kubenswrapper[4799]: I1129 04:38:45.669029 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"4e75dbd2828109703f688ee79c9b4e1070338397acb59de75002b9a6155141f9"} Nov 29 04:38:45 crc kubenswrapper[4799]: W1129 04:38:45.743576 4799 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.83:6443: connect: connection refused Nov 29 04:38:45 crc kubenswrapper[4799]: E1129 04:38:45.743647 4799 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.83:6443: connect: connection refused" logger="UnhandledError" Nov 29 04:38:46 crc kubenswrapper[4799]: E1129 04:38:46.004412 4799 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.83:6443: connect: connection refused" interval="1.6s" Nov 29 04:38:46 crc kubenswrapper[4799]: W1129 04:38:46.143847 4799 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.83:6443: connect: connection refused Nov 29 04:38:46 crc kubenswrapper[4799]: E1129 04:38:46.143945 4799 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.83:6443: connect: connection refused" logger="UnhandledError" Nov 29 04:38:46 crc kubenswrapper[4799]: W1129 04:38:46.244288 4799 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.83:6443: connect: connection refused Nov 29 04:38:46 crc kubenswrapper[4799]: E1129 04:38:46.244439 4799 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.83:6443: connect: connection refused" logger="UnhandledError" Nov 29 04:38:46 crc kubenswrapper[4799]: I1129 04:38:46.285251 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:38:46 crc kubenswrapper[4799]: I1129 04:38:46.286458 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:38:46 crc kubenswrapper[4799]: I1129 04:38:46.286500 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:38:46 crc kubenswrapper[4799]: I1129 04:38:46.286513 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:38:46 crc kubenswrapper[4799]: I1129 04:38:46.286538 4799 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 29 04:38:46 crc kubenswrapper[4799]: E1129 04:38:46.287005 4799 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.83:6443: connect: connection refused" node="crc" Nov 29 04:38:46 crc kubenswrapper[4799]: I1129 04:38:46.599941 4799 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.83:6443: connect: connection refused Nov 29 04:38:46 crc kubenswrapper[4799]: I1129 04:38:46.641199 4799 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Nov 29 04:38:46 crc kubenswrapper[4799]: E1129 04:38:46.642083 4799 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.83:6443: connect: connection refused" logger="UnhandledError" Nov 29 04:38:46 crc kubenswrapper[4799]: I1129 04:38:46.672599 4799 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="3594d8872daaa452e159a330f34c8d9f3ebfd44d812778c341e88061deef7668" exitCode=0 Nov 29 04:38:46 crc kubenswrapper[4799]: I1129 04:38:46.672686 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:38:46 crc kubenswrapper[4799]: I1129 04:38:46.672762 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"3594d8872daaa452e159a330f34c8d9f3ebfd44d812778c341e88061deef7668"} Nov 29 04:38:46 crc kubenswrapper[4799]: I1129 04:38:46.675893 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:38:46 crc kubenswrapper[4799]: I1129 04:38:46.675941 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:38:46 crc kubenswrapper[4799]: I1129 04:38:46.675961 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:38:46 crc kubenswrapper[4799]: I1129 04:38:46.677326 4799 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="8d5b8064433d389b05e3e4744ef4c1f72512db615eae94e2a847bbd4cb2518f3" exitCode=0 Nov 29 04:38:46 crc kubenswrapper[4799]: I1129 04:38:46.677428 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"8d5b8064433d389b05e3e4744ef4c1f72512db615eae94e2a847bbd4cb2518f3"} Nov 29 04:38:46 crc kubenswrapper[4799]: I1129 04:38:46.677575 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:38:46 crc kubenswrapper[4799]: I1129 04:38:46.678828 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:38:46 crc kubenswrapper[4799]: I1129 04:38:46.678862 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:38:46 crc kubenswrapper[4799]: I1129 04:38:46.678875 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:38:46 crc kubenswrapper[4799]: I1129 04:38:46.682099 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"3ebed9ec415bb433935da36df8014f9087236cea9788b2048cbff8714f8ad0df"} Nov 29 04:38:46 crc kubenswrapper[4799]: I1129 04:38:46.682119 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:38:46 crc kubenswrapper[4799]: I1129 04:38:46.682131 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"16f42a55d797c6c94af3a5eac1bf08aa38a82359b7ddd0ba32dcc797218ceee6"} Nov 29 04:38:46 crc kubenswrapper[4799]: I1129 04:38:46.682146 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"33c0540ccb097b15f65369ff537aef4ae621cc828ba314eb61f0a0037608cadf"} Nov 29 04:38:46 crc kubenswrapper[4799]: I1129 04:38:46.682158 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ea4c241d29297a73dce12bd0bc2863c56475eb35b8fe8c93991ba0ef8695e38f"} Nov 29 04:38:46 crc kubenswrapper[4799]: I1129 04:38:46.683241 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:38:46 crc kubenswrapper[4799]: I1129 04:38:46.683265 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:38:46 crc kubenswrapper[4799]: I1129 04:38:46.683274 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:38:46 crc kubenswrapper[4799]: I1129 04:38:46.683566 4799 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f" exitCode=0 Nov 29 04:38:46 crc kubenswrapper[4799]: I1129 04:38:46.683619 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f"} Nov 29 04:38:46 crc kubenswrapper[4799]: I1129 04:38:46.683696 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:38:46 crc kubenswrapper[4799]: I1129 04:38:46.684388 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:38:46 crc kubenswrapper[4799]: I1129 04:38:46.684414 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:38:46 crc kubenswrapper[4799]: I1129 04:38:46.684460 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:38:46 crc kubenswrapper[4799]: I1129 04:38:46.685590 4799 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="9b2129acb1f879e6e104b392aa6699d6bcf8b3c373bff0e17a2aa2e658c700da" exitCode=0 Nov 29 04:38:46 crc kubenswrapper[4799]: I1129 04:38:46.685716 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:38:46 crc kubenswrapper[4799]: I1129 04:38:46.686099 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"9b2129acb1f879e6e104b392aa6699d6bcf8b3c373bff0e17a2aa2e658c700da"} Nov 29 04:38:46 crc kubenswrapper[4799]: I1129 04:38:46.686364 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:38:46 crc kubenswrapper[4799]: I1129 04:38:46.686387 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:38:46 crc kubenswrapper[4799]: I1129 04:38:46.686395 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:38:46 crc kubenswrapper[4799]: I1129 04:38:46.687387 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:38:46 crc kubenswrapper[4799]: I1129 04:38:46.687946 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:38:46 crc kubenswrapper[4799]: I1129 04:38:46.687968 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:38:46 crc kubenswrapper[4799]: I1129 04:38:46.687977 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:38:47 crc kubenswrapper[4799]: I1129 04:38:47.065345 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 29 04:38:47 crc kubenswrapper[4799]: I1129 04:38:47.690570 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"624c4a09beeecd568d6045a720dfa28c15821d931021cedd8776da4bb36b557f"} Nov 29 04:38:47 crc kubenswrapper[4799]: I1129 04:38:47.690892 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"a02c75c8691748a626cbef2de964cb1f4579e702276e8b02d805db0876a8fee9"} Nov 29 04:38:47 crc kubenswrapper[4799]: I1129 04:38:47.690904 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"186c216d29cdc56583cd5f39acc7cfd0558f60c668b1f5cbf039b566c1747e12"} Nov 29 04:38:47 crc kubenswrapper[4799]: I1129 04:38:47.690664 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:38:47 crc kubenswrapper[4799]: I1129 04:38:47.691669 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:38:47 crc kubenswrapper[4799]: I1129 04:38:47.691724 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:38:47 crc kubenswrapper[4799]: I1129 04:38:47.691742 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:38:47 crc kubenswrapper[4799]: I1129 04:38:47.694549 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d7762237003a5a256ff20a16a4869462e2e0dd1426aa391f7ec74ed9b1f572d4"} Nov 29 04:38:47 crc kubenswrapper[4799]: I1129 04:38:47.694595 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc"} Nov 29 04:38:47 crc kubenswrapper[4799]: I1129 04:38:47.694618 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2"} Nov 29 04:38:47 crc kubenswrapper[4799]: I1129 04:38:47.694630 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe"} Nov 29 04:38:47 crc kubenswrapper[4799]: I1129 04:38:47.694650 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea"} Nov 29 04:38:47 crc kubenswrapper[4799]: I1129 04:38:47.694628 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:38:47 crc kubenswrapper[4799]: I1129 04:38:47.695543 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:38:47 crc kubenswrapper[4799]: I1129 04:38:47.695586 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:38:47 crc kubenswrapper[4799]: I1129 04:38:47.695603 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:38:47 crc kubenswrapper[4799]: I1129 04:38:47.696533 4799 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="cefc47f4d22cbadb998e1591ab51bf1c861affec5facdf99f7f622394b69befb" exitCode=0 Nov 29 04:38:47 crc kubenswrapper[4799]: I1129 04:38:47.696587 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"cefc47f4d22cbadb998e1591ab51bf1c861affec5facdf99f7f622394b69befb"} Nov 29 04:38:47 crc kubenswrapper[4799]: I1129 04:38:47.696662 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:38:47 crc kubenswrapper[4799]: I1129 04:38:47.697574 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:38:47 crc kubenswrapper[4799]: I1129 04:38:47.697617 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:38:47 crc kubenswrapper[4799]: I1129 04:38:47.697633 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:38:47 crc kubenswrapper[4799]: I1129 04:38:47.698232 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"358995762bd89ae2b4455ece3b30fef5e04c6439899796940d99c5ab731612eb"} Nov 29 04:38:47 crc kubenswrapper[4799]: I1129 04:38:47.698254 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:38:47 crc kubenswrapper[4799]: I1129 04:38:47.698274 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:38:47 crc kubenswrapper[4799]: I1129 04:38:47.699065 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:38:47 crc kubenswrapper[4799]: I1129 04:38:47.699078 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:38:47 crc kubenswrapper[4799]: I1129 04:38:47.699095 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:38:47 crc kubenswrapper[4799]: I1129 04:38:47.699105 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:38:47 crc kubenswrapper[4799]: I1129 04:38:47.699096 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:38:47 crc kubenswrapper[4799]: I1129 04:38:47.699142 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:38:47 crc kubenswrapper[4799]: I1129 04:38:47.887578 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:38:47 crc kubenswrapper[4799]: I1129 04:38:47.888534 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:38:47 crc kubenswrapper[4799]: I1129 04:38:47.888562 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:38:47 crc kubenswrapper[4799]: I1129 04:38:47.888571 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:38:47 crc kubenswrapper[4799]: I1129 04:38:47.888590 4799 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 29 04:38:48 crc kubenswrapper[4799]: I1129 04:38:48.264126 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 29 04:38:48 crc kubenswrapper[4799]: I1129 04:38:48.704934 4799 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="9e113c9d79d9e1d5336a66670c7891f337f4518fca018eee02481d2a6503b26c" exitCode=0 Nov 29 04:38:48 crc kubenswrapper[4799]: I1129 04:38:48.704994 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"9e113c9d79d9e1d5336a66670c7891f337f4518fca018eee02481d2a6503b26c"} Nov 29 04:38:48 crc kubenswrapper[4799]: I1129 04:38:48.705070 4799 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 29 04:38:48 crc kubenswrapper[4799]: I1129 04:38:48.705107 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:38:48 crc kubenswrapper[4799]: I1129 04:38:48.705160 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:38:48 crc kubenswrapper[4799]: I1129 04:38:48.705230 4799 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 29 04:38:48 crc kubenswrapper[4799]: I1129 04:38:48.705255 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:38:48 crc kubenswrapper[4799]: I1129 04:38:48.705116 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:38:48 crc kubenswrapper[4799]: I1129 04:38:48.705161 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:38:48 crc kubenswrapper[4799]: I1129 04:38:48.707015 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:38:48 crc kubenswrapper[4799]: I1129 04:38:48.707039 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:38:48 crc kubenswrapper[4799]: I1129 04:38:48.707044 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:38:48 crc kubenswrapper[4799]: I1129 04:38:48.707062 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:38:48 crc kubenswrapper[4799]: I1129 04:38:48.707091 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:38:48 crc kubenswrapper[4799]: I1129 04:38:48.707120 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:38:48 crc kubenswrapper[4799]: I1129 04:38:48.707141 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:38:48 crc kubenswrapper[4799]: I1129 04:38:48.707061 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:38:48 crc kubenswrapper[4799]: I1129 04:38:48.707268 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:38:48 crc kubenswrapper[4799]: I1129 04:38:48.707835 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:38:48 crc kubenswrapper[4799]: I1129 04:38:48.707895 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:38:48 crc kubenswrapper[4799]: I1129 04:38:48.707919 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:38:48 crc kubenswrapper[4799]: I1129 04:38:48.708161 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:38:48 crc kubenswrapper[4799]: I1129 04:38:48.708220 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:38:48 crc kubenswrapper[4799]: I1129 04:38:48.708245 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:38:48 crc kubenswrapper[4799]: I1129 04:38:48.961075 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 29 04:38:49 crc kubenswrapper[4799]: I1129 04:38:49.713609 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c5ed3b1ac4ea270da98bb14639a0d51e1aa5eacaa289f1ec91334a71acaef2ed"} Nov 29 04:38:49 crc kubenswrapper[4799]: I1129 04:38:49.713662 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"169a72781249757026a51efe63174203b02d9f04a1cf42cafa852529f4e2980d"} Nov 29 04:38:49 crc kubenswrapper[4799]: I1129 04:38:49.713677 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"fd3cb161e0e04a1fb4255212bff5ffd70fa74b92ab0cddc1fc5e749199ad0706"} Nov 29 04:38:49 crc kubenswrapper[4799]: I1129 04:38:49.713689 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"074beebad56cc8402a153545bd158fa778aa0acc805b048fc0ad85afd9dcaa11"} Nov 29 04:38:49 crc kubenswrapper[4799]: I1129 04:38:49.713705 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3a6b9e8b658317e4c69c986b5d20d296d3c746315b73b5f8277bfbb5acf57307"} Nov 29 04:38:49 crc kubenswrapper[4799]: I1129 04:38:49.713711 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:38:49 crc kubenswrapper[4799]: I1129 04:38:49.713711 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:38:49 crc kubenswrapper[4799]: I1129 04:38:49.714881 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:38:49 crc kubenswrapper[4799]: I1129 04:38:49.714909 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:38:49 crc kubenswrapper[4799]: I1129 04:38:49.714920 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:38:49 crc kubenswrapper[4799]: I1129 04:38:49.715668 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:38:49 crc kubenswrapper[4799]: I1129 04:38:49.715702 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:38:49 crc kubenswrapper[4799]: I1129 04:38:49.715713 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:38:50 crc kubenswrapper[4799]: I1129 04:38:50.025900 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 04:38:50 crc kubenswrapper[4799]: I1129 04:38:50.026062 4799 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 29 04:38:50 crc kubenswrapper[4799]: I1129 04:38:50.026103 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:38:50 crc kubenswrapper[4799]: I1129 04:38:50.027187 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:38:50 crc kubenswrapper[4799]: I1129 04:38:50.027326 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:38:50 crc kubenswrapper[4799]: I1129 04:38:50.027447 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:38:50 crc kubenswrapper[4799]: I1129 04:38:50.716284 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:38:50 crc kubenswrapper[4799]: I1129 04:38:50.716874 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:38:50 crc kubenswrapper[4799]: I1129 04:38:50.717358 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:38:50 crc kubenswrapper[4799]: I1129 04:38:50.717391 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:38:50 crc kubenswrapper[4799]: I1129 04:38:50.717402 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:38:50 crc kubenswrapper[4799]: I1129 04:38:50.718221 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:38:50 crc kubenswrapper[4799]: I1129 04:38:50.718249 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:38:50 crc kubenswrapper[4799]: I1129 04:38:50.718257 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:38:50 crc kubenswrapper[4799]: I1129 04:38:50.780200 4799 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Nov 29 04:38:51 crc kubenswrapper[4799]: I1129 04:38:51.397151 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Nov 29 04:38:51 crc kubenswrapper[4799]: I1129 04:38:51.718829 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:38:51 crc kubenswrapper[4799]: I1129 04:38:51.720033 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:38:51 crc kubenswrapper[4799]: I1129 04:38:51.720094 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:38:51 crc kubenswrapper[4799]: I1129 04:38:51.720111 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:38:51 crc kubenswrapper[4799]: I1129 04:38:51.951133 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 04:38:51 crc kubenswrapper[4799]: I1129 04:38:51.951307 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:38:51 crc kubenswrapper[4799]: I1129 04:38:51.952726 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:38:51 crc kubenswrapper[4799]: I1129 04:38:51.952758 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:38:51 crc kubenswrapper[4799]: I1129 04:38:51.952767 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:38:51 crc kubenswrapper[4799]: I1129 04:38:51.961480 4799 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 29 04:38:51 crc kubenswrapper[4799]: I1129 04:38:51.961537 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 29 04:38:52 crc kubenswrapper[4799]: I1129 04:38:52.036630 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 04:38:52 crc kubenswrapper[4799]: I1129 04:38:52.738631 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:38:52 crc kubenswrapper[4799]: I1129 04:38:52.739993 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:38:52 crc kubenswrapper[4799]: I1129 04:38:52.740035 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:38:52 crc kubenswrapper[4799]: I1129 04:38:52.740046 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:38:52 crc kubenswrapper[4799]: I1129 04:38:52.770287 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 29 04:38:52 crc kubenswrapper[4799]: I1129 04:38:52.770484 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:38:52 crc kubenswrapper[4799]: I1129 04:38:52.771477 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:38:52 crc kubenswrapper[4799]: I1129 04:38:52.771504 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:38:52 crc kubenswrapper[4799]: I1129 04:38:52.771512 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:38:53 crc kubenswrapper[4799]: I1129 04:38:53.724310 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Nov 29 04:38:53 crc kubenswrapper[4799]: I1129 04:38:53.724647 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:38:53 crc kubenswrapper[4799]: I1129 04:38:53.730750 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:38:53 crc kubenswrapper[4799]: I1129 04:38:53.730927 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:38:53 crc kubenswrapper[4799]: I1129 04:38:53.731027 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:38:54 crc kubenswrapper[4799]: I1129 04:38:54.246968 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 29 04:38:54 crc kubenswrapper[4799]: I1129 04:38:54.247162 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:38:54 crc kubenswrapper[4799]: I1129 04:38:54.248737 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:38:54 crc kubenswrapper[4799]: I1129 04:38:54.248853 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:38:54 crc kubenswrapper[4799]: I1129 04:38:54.248869 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:38:54 crc kubenswrapper[4799]: I1129 04:38:54.253071 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 29 04:38:54 crc kubenswrapper[4799]: I1129 04:38:54.744682 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:38:54 crc kubenswrapper[4799]: I1129 04:38:54.746373 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:38:54 crc kubenswrapper[4799]: I1129 04:38:54.746436 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:38:54 crc kubenswrapper[4799]: I1129 04:38:54.746452 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:38:54 crc kubenswrapper[4799]: E1129 04:38:54.786154 4799 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 29 04:38:57 crc kubenswrapper[4799]: W1129 04:38:57.259014 4799 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Nov 29 04:38:57 crc kubenswrapper[4799]: I1129 04:38:57.259126 4799 trace.go:236] Trace[1789366372]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (29-Nov-2025 04:38:47.257) (total time: 10001ms): Nov 29 04:38:57 crc kubenswrapper[4799]: Trace[1789366372]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (04:38:57.258) Nov 29 04:38:57 crc kubenswrapper[4799]: Trace[1789366372]: [10.001314875s] [10.001314875s] END Nov 29 04:38:57 crc kubenswrapper[4799]: E1129 04:38:57.259151 4799 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Nov 29 04:38:57 crc kubenswrapper[4799]: W1129 04:38:57.428865 4799 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Nov 29 04:38:57 crc kubenswrapper[4799]: I1129 04:38:57.428960 4799 trace.go:236] Trace[1499052837]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (29-Nov-2025 04:38:47.427) (total time: 10001ms): Nov 29 04:38:57 crc kubenswrapper[4799]: Trace[1499052837]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (04:38:57.428) Nov 29 04:38:57 crc kubenswrapper[4799]: Trace[1499052837]: [10.001472964s] [10.001472964s] END Nov 29 04:38:57 crc kubenswrapper[4799]: E1129 04:38:57.428984 4799 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Nov 29 04:38:57 crc kubenswrapper[4799]: I1129 04:38:57.599826 4799 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Nov 29 04:38:57 crc kubenswrapper[4799]: E1129 04:38:57.605011 4799 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="3.2s" Nov 29 04:38:57 crc kubenswrapper[4799]: E1129 04:38:57.890116 4799 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="crc" Nov 29 04:38:57 crc kubenswrapper[4799]: I1129 04:38:57.898637 4799 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 29 04:38:57 crc kubenswrapper[4799]: I1129 04:38:57.898688 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 29 04:38:57 crc kubenswrapper[4799]: I1129 04:38:57.904165 4799 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 29 04:38:57 crc kubenswrapper[4799]: I1129 04:38:57.904249 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 29 04:38:58 crc kubenswrapper[4799]: I1129 04:38:58.268816 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 29 04:38:58 crc kubenswrapper[4799]: I1129 04:38:58.268931 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:38:58 crc kubenswrapper[4799]: I1129 04:38:58.270125 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:38:58 crc kubenswrapper[4799]: I1129 04:38:58.270163 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:38:58 crc kubenswrapper[4799]: I1129 04:38:58.270177 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:00 crc kubenswrapper[4799]: I1129 04:39:00.033145 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 04:39:00 crc kubenswrapper[4799]: I1129 04:39:00.033312 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:39:00 crc kubenswrapper[4799]: I1129 04:39:00.033565 4799 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 29 04:39:00 crc kubenswrapper[4799]: I1129 04:39:00.033595 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 29 04:39:00 crc kubenswrapper[4799]: I1129 04:39:00.034319 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:00 crc kubenswrapper[4799]: I1129 04:39:00.034343 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:00 crc kubenswrapper[4799]: I1129 04:39:00.034354 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:00 crc kubenswrapper[4799]: I1129 04:39:00.041604 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 04:39:00 crc kubenswrapper[4799]: I1129 04:39:00.758972 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:39:00 crc kubenswrapper[4799]: I1129 04:39:00.759503 4799 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 29 04:39:00 crc kubenswrapper[4799]: I1129 04:39:00.759556 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 29 04:39:00 crc kubenswrapper[4799]: I1129 04:39:00.759782 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:00 crc kubenswrapper[4799]: I1129 04:39:00.759872 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:00 crc kubenswrapper[4799]: I1129 04:39:00.759899 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:01 crc kubenswrapper[4799]: I1129 04:39:01.091210 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:39:01 crc kubenswrapper[4799]: I1129 04:39:01.092297 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:01 crc kubenswrapper[4799]: I1129 04:39:01.092351 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:01 crc kubenswrapper[4799]: I1129 04:39:01.092368 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:01 crc kubenswrapper[4799]: I1129 04:39:01.092406 4799 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 29 04:39:01 crc kubenswrapper[4799]: E1129 04:39:01.095945 4799 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Nov 29 04:39:01 crc kubenswrapper[4799]: I1129 04:39:01.426553 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Nov 29 04:39:01 crc kubenswrapper[4799]: I1129 04:39:01.426755 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:39:01 crc kubenswrapper[4799]: I1129 04:39:01.428241 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:01 crc kubenswrapper[4799]: I1129 04:39:01.428290 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:01 crc kubenswrapper[4799]: I1129 04:39:01.428307 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:01 crc kubenswrapper[4799]: I1129 04:39:01.437908 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Nov 29 04:39:01 crc kubenswrapper[4799]: I1129 04:39:01.762005 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:39:01 crc kubenswrapper[4799]: I1129 04:39:01.763671 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:01 crc kubenswrapper[4799]: I1129 04:39:01.763752 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:01 crc kubenswrapper[4799]: I1129 04:39:01.763764 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:01 crc kubenswrapper[4799]: I1129 04:39:01.951963 4799 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 29 04:39:01 crc kubenswrapper[4799]: I1129 04:39:01.952036 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 29 04:39:01 crc kubenswrapper[4799]: I1129 04:39:01.958469 4799 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 29 04:39:01 crc kubenswrapper[4799]: I1129 04:39:01.958559 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 29 04:39:01 crc kubenswrapper[4799]: I1129 04:39:01.962582 4799 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 29 04:39:01 crc kubenswrapper[4799]: I1129 04:39:01.962629 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 29 04:39:02 crc kubenswrapper[4799]: I1129 04:39:02.272298 4799 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 29 04:39:02 crc kubenswrapper[4799]: I1129 04:39:02.902838 4799 trace.go:236] Trace[1421504224]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (29-Nov-2025 04:38:48.526) (total time: 14376ms): Nov 29 04:39:02 crc kubenswrapper[4799]: Trace[1421504224]: ---"Objects listed" error: 14376ms (04:39:02.902) Nov 29 04:39:02 crc kubenswrapper[4799]: Trace[1421504224]: [14.376636823s] [14.376636823s] END Nov 29 04:39:02 crc kubenswrapper[4799]: I1129 04:39:02.902875 4799 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 29 04:39:02 crc kubenswrapper[4799]: I1129 04:39:02.902970 4799 trace.go:236] Trace[587692900]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (29-Nov-2025 04:38:49.336) (total time: 13565ms): Nov 29 04:39:02 crc kubenswrapper[4799]: Trace[587692900]: ---"Objects listed" error: 13565ms (04:39:02.902) Nov 29 04:39:02 crc kubenswrapper[4799]: Trace[587692900]: [13.565976322s] [13.565976322s] END Nov 29 04:39:02 crc kubenswrapper[4799]: I1129 04:39:02.902990 4799 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 29 04:39:02 crc kubenswrapper[4799]: I1129 04:39:02.903064 4799 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Nov 29 04:39:02 crc kubenswrapper[4799]: I1129 04:39:02.907804 4799 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.457659 4799 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.743302 4799 apiserver.go:52] "Watching apiserver" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.746586 4799 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.746973 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.747388 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.747521 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.747583 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 04:39:03 crc kubenswrapper[4799]: E1129 04:39:03.747635 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 04:39:03 crc kubenswrapper[4799]: E1129 04:39:03.747654 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.747827 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.747923 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.748533 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:39:03 crc kubenswrapper[4799]: E1129 04:39:03.748593 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.750481 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.750606 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.750635 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.750678 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.750706 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.751329 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.751467 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.751573 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.751665 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.768016 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.769980 4799 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d7762237003a5a256ff20a16a4869462e2e0dd1426aa391f7ec74ed9b1f572d4" exitCode=255 Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.770033 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"d7762237003a5a256ff20a16a4869462e2e0dd1426aa391f7ec74ed9b1f572d4"} Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.781424 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.781445 4799 scope.go:117] "RemoveContainer" containerID="d7762237003a5a256ff20a16a4869462e2e0dd1426aa391f7ec74ed9b1f572d4" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.786570 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.801525 4799 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.803146 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.808730 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.808778 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.808820 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.808841 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.808858 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.808878 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.808894 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.808914 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.808934 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.808950 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.808967 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.808986 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.809002 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.809018 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.809042 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.809058 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.809072 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.809090 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.809105 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.809122 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.809138 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.809155 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.809170 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.809208 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.809225 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.809241 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.809258 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.809277 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.809293 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.809307 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.809322 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.809337 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.809353 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.809371 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.809391 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.809409 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.809424 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.809439 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.809454 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.809470 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.809485 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.809503 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.809496 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.809518 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.809671 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.809712 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.809771 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.809815 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.809840 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.809825 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.809860 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.809882 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.809903 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.809922 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.809942 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.809961 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.809983 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810000 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810017 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810032 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810050 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810066 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810082 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810098 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810115 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810131 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810149 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810163 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810178 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810196 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810212 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810228 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810244 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810259 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810278 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810293 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810313 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810328 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810343 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810357 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810374 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810391 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810407 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810423 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810441 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810456 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810482 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810497 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810511 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810526 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810548 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810569 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810591 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810607 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810622 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810638 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810657 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810672 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810688 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810703 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810721 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810740 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810758 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810774 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810811 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810841 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810858 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810876 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810897 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810915 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810932 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810950 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810966 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810983 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810999 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811014 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811032 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811047 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811138 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811157 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811174 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811191 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811206 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811222 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811239 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811258 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811275 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811291 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811308 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811326 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811344 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811360 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811378 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811395 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811412 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811429 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811446 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811463 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811479 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811496 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811512 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811529 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811547 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811565 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811588 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811605 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811624 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811641 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811660 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811678 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811697 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811714 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811732 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811748 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811765 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811782 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811813 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811829 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811845 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811862 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811878 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811897 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811916 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811933 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811952 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811970 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811988 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.812005 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.812023 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.812041 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.812061 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.812098 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.812116 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.812133 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.812149 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.812167 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.812183 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.812201 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.812222 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.812238 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.812257 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.812273 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.812291 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.812310 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.812328 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.812344 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.812361 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.812379 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.812398 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.812418 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.812435 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.812453 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.812470 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.812487 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.812504 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.812521 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.812708 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.812731 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.812751 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.812818 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.812848 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.812870 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.812889 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.812908 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.812929 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.812948 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.812969 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.812992 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.813011 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.813031 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.813051 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.813081 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.813102 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.813152 4799 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.813168 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.813178 4799 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.813187 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.809959 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810102 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810296 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810723 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810831 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810833 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.810878 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811338 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811545 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811653 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811952 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.811953 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.812013 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.812064 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.812336 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.812543 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.812548 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.812709 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.812860 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.813955 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.812896 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.813204 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.814058 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.813062 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.814111 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.813404 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.813556 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.814168 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.813620 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.813949 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.814079 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.814093 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.813555 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.814397 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.814502 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.814597 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.814878 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.815109 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.815325 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.815476 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.815653 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.815779 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.815765 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.816529 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.816757 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.816937 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.816965 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.817459 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.817480 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.817825 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.817872 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.818097 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.818268 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.818648 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.818673 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.818769 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.818846 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.818912 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.819170 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.819186 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.819737 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.819692 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.819832 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.819911 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.820106 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.820196 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.820434 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.820458 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.820489 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.820470 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.820730 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.820757 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.820841 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.821062 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.821082 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.821124 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.821377 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.821396 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.821541 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.821588 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.821680 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.821680 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.821884 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.821934 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.822187 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.822459 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.822368 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.821920 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.822543 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.822530 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.822957 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.823550 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.823676 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.823771 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.823814 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.824362 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.824434 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.824683 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.825163 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.826084 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.826296 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.826387 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.826407 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.826422 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.826523 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.826558 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.826733 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.826746 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.827399 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.826470 4799 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.827849 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 29 04:39:03 crc kubenswrapper[4799]: E1129 04:39:03.828072 4799 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 29 04:39:03 crc kubenswrapper[4799]: E1129 04:39:03.828180 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-29 04:39:04.328147053 +0000 UTC m=+19.971077503 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 29 04:39:03 crc kubenswrapper[4799]: E1129 04:39:03.828511 4799 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 29 04:39:03 crc kubenswrapper[4799]: E1129 04:39:03.828602 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-29 04:39:04.328576044 +0000 UTC m=+19.971506504 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.828674 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.829028 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.829708 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: E1129 04:39:03.829749 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 04:39:04.329725762 +0000 UTC m=+19.972656202 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.829995 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.830096 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.830683 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.830798 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.831165 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.831195 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.831641 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.831827 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.831933 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.832082 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.832349 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.832378 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.832538 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.832548 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.832618 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.832680 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.832704 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.832943 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.833074 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.833176 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.833319 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.833945 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.834132 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.834576 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.835258 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.835313 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.836165 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.836379 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: E1129 04:39:03.836498 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.836499 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: E1129 04:39:03.836518 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 29 04:39:03 crc kubenswrapper[4799]: E1129 04:39:03.836533 4799 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 04:39:03 crc kubenswrapper[4799]: E1129 04:39:03.836591 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-29 04:39:04.336571651 +0000 UTC m=+19.979502061 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.836915 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.837275 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.837780 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.838190 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.838738 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.839911 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.839965 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.842509 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.842616 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.843074 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.843485 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.843767 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.844511 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.844877 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.846303 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.847018 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.847135 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.847314 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.847478 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: E1129 04:39:03.847740 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 29 04:39:03 crc kubenswrapper[4799]: E1129 04:39:03.847766 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 29 04:39:03 crc kubenswrapper[4799]: E1129 04:39:03.847778 4799 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 04:39:03 crc kubenswrapper[4799]: E1129 04:39:03.847850 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-29 04:39:04.347832668 +0000 UTC m=+19.990763068 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.848093 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.848245 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.849095 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.849176 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.849174 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.849339 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.849587 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.850195 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.850328 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.850565 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.850603 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.851428 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.851583 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.851725 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.854776 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.854950 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.854955 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.855089 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.855372 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.855580 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.855953 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.856429 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.859281 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.859665 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.859712 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.859772 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.860818 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.861141 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.861214 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.861244 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.861716 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.861866 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.862223 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.862562 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.863232 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.866071 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.868454 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d40160e-075f-44b7-9d9d-3291e7774f27\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7762237003a5a256ff20a16a4869462e2e0dd1426aa391f7ec74ed9b1f572d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7762237003a5a256ff20a16a4869462e2e0dd1426aa391f7ec74ed9b1f572d4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1129 04:39:02.908489 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1129 04:39:02.908637 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 04:39:02.909677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703886220/tls.crt::/tmp/serving-cert-1703886220/tls.key\\\\\\\"\\\\nI1129 04:39:03.077876 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 04:39:03.080368 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 04:39:03.080390 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 04:39:03.080424 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 04:39:03.080432 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 04:39:03.086120 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 04:39:03.086240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086412 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 04:39:03.086502 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 04:39:03.086563 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 04:39:03.086622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1129 04:39:03.086141 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1129 04:39:03.086836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:38:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.877441 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.878489 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.883051 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.887074 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.889706 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.893883 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.900887 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.912589 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914078 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914147 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914270 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914440 4799 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914473 4799 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914489 4799 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914504 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914521 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914546 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914559 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914550 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914578 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914631 4799 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914641 4799 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914653 4799 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914664 4799 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914672 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914682 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914692 4799 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914718 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914728 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914737 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914746 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914755 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914763 4799 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914772 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914780 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914802 4799 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914812 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914820 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914829 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914837 4799 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914845 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914853 4799 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914861 4799 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914869 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914879 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914887 4799 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914895 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914903 4799 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914912 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914921 4799 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914930 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914939 4799 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914947 4799 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914955 4799 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914963 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914972 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914980 4799 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914989 4799 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.914997 4799 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915006 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915014 4799 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915023 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915034 4799 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915055 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915064 4799 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915072 4799 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915099 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915108 4799 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915116 4799 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915125 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915133 4799 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915142 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915150 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915158 4799 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915166 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915174 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915187 4799 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915195 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915203 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915212 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915220 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915228 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915236 4799 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915246 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915254 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915262 4799 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915275 4799 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915283 4799 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915291 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915299 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915307 4799 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915314 4799 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915323 4799 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915330 4799 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915338 4799 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915347 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915355 4799 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915364 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915371 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915391 4799 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915399 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915412 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915421 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915430 4799 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915441 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915449 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915457 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915465 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915474 4799 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915650 4799 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915704 4799 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915713 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915721 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915729 4799 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915766 4799 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915774 4799 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915781 4799 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915802 4799 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915810 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915818 4799 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915826 4799 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915834 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915841 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915849 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915857 4799 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915883 4799 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915891 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915901 4799 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915909 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915916 4799 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915924 4799 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915932 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915942 4799 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915951 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915970 4799 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915985 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.915993 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916003 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916012 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916020 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916030 4799 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916038 4799 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916047 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916056 4799 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916064 4799 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916072 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916080 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916088 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916098 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916107 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916117 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916125 4799 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916133 4799 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916144 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916152 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916160 4799 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916167 4799 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916179 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916188 4799 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916199 4799 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916207 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916214 4799 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916222 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916230 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916238 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916250 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916258 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916265 4799 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916273 4799 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916280 4799 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916294 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916319 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916330 4799 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916337 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916345 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916352 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916360 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916367 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916375 4799 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916382 4799 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916392 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916401 4799 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916409 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916418 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916435 4799 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916448 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916465 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916474 4799 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916482 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916490 4799 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916498 4799 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916518 4799 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916525 4799 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916533 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916540 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916548 4799 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916558 4799 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916568 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916582 4799 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916590 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916599 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:03 crc kubenswrapper[4799]: I1129 04:39:03.916607 4799 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.059864 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.066779 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 29 04:39:04 crc kubenswrapper[4799]: W1129 04:39:04.071932 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-2e2b783d14d219076ec2f389ad2ff4e6d42a1c7ac369ce3dc23db350f24e7ede WatchSource:0}: Error finding container 2e2b783d14d219076ec2f389ad2ff4e6d42a1c7ac369ce3dc23db350f24e7ede: Status 404 returned error can't find the container with id 2e2b783d14d219076ec2f389ad2ff4e6d42a1c7ac369ce3dc23db350f24e7ede Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.073998 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 29 04:39:04 crc kubenswrapper[4799]: W1129 04:39:04.087305 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-505e78ddf4f09331c2388f1fa7083f24e2dc672f77671a10eec53d6ba4bb1bef WatchSource:0}: Error finding container 505e78ddf4f09331c2388f1fa7083f24e2dc672f77671a10eec53d6ba4bb1bef: Status 404 returned error can't find the container with id 505e78ddf4f09331c2388f1fa7083f24e2dc672f77671a10eec53d6ba4bb1bef Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.419964 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.420044 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.420077 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.420101 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.420128 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 04:39:04 crc kubenswrapper[4799]: E1129 04:39:04.420228 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 04:39:05.420188239 +0000 UTC m=+21.063118669 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:39:04 crc kubenswrapper[4799]: E1129 04:39:04.420257 4799 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 29 04:39:04 crc kubenswrapper[4799]: E1129 04:39:04.420248 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 29 04:39:04 crc kubenswrapper[4799]: E1129 04:39:04.420364 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 29 04:39:04 crc kubenswrapper[4799]: E1129 04:39:04.420401 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 29 04:39:04 crc kubenswrapper[4799]: E1129 04:39:04.420414 4799 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 04:39:04 crc kubenswrapper[4799]: E1129 04:39:04.420378 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 29 04:39:04 crc kubenswrapper[4799]: E1129 04:39:04.420471 4799 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 04:39:04 crc kubenswrapper[4799]: E1129 04:39:04.420340 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-29 04:39:05.420321382 +0000 UTC m=+21.063251772 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 29 04:39:04 crc kubenswrapper[4799]: E1129 04:39:04.420509 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-29 04:39:05.420493476 +0000 UTC m=+21.063423876 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 04:39:04 crc kubenswrapper[4799]: E1129 04:39:04.420521 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-29 04:39:05.420515997 +0000 UTC m=+21.063446397 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 04:39:04 crc kubenswrapper[4799]: E1129 04:39:04.420259 4799 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 29 04:39:04 crc kubenswrapper[4799]: E1129 04:39:04.420616 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-29 04:39:05.420593699 +0000 UTC m=+21.063524249 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.661951 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.663192 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.664518 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.665313 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.666885 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.667569 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.668848 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.669911 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.671197 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.672038 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.673159 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.674122 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.675311 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.676001 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.676774 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.678209 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.678995 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.679697 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.680390 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.680960 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.681413 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.682007 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.682422 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.683032 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.683405 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.684007 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.684626 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.686344 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.687254 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.688086 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.688753 4799 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.688764 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:04Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.688925 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.690817 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.691601 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.693200 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.695168 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.695969 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.696611 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.697412 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.698266 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.698907 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.699630 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.702269 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.702319 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:04Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.703355 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.703811 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.704780 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.705313 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.706401 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.706858 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.707619 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.708071 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.708535 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.709487 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.709941 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.715245 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:04Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.733968 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:04Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.746800 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d40160e-075f-44b7-9d9d-3291e7774f27\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7762237003a5a256ff20a16a4869462e2e0dd1426aa391f7ec74ed9b1f572d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7762237003a5a256ff20a16a4869462e2e0dd1426aa391f7ec74ed9b1f572d4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1129 04:39:02.908489 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1129 04:39:02.908637 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 04:39:02.909677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703886220/tls.crt::/tmp/serving-cert-1703886220/tls.key\\\\\\\"\\\\nI1129 04:39:03.077876 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 04:39:03.080368 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 04:39:03.080390 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 04:39:03.080424 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 04:39:03.080432 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 04:39:03.086120 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 04:39:03.086240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086412 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 04:39:03.086502 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 04:39:03.086563 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 04:39:03.086622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1129 04:39:03.086141 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1129 04:39:03.086836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:38:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:04Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.759280 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:04Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.774039 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:04Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.775067 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"ee17b4a9a8999e5f76ceaf69dbd13acf80ed0e95e72f3e1428ce38c5d5982788"} Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.775113 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"382256e0fa1a4c26610227b9ae96288ccbfbbd97aa3cbd5fd4f18cb4228a234b"} Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.775127 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"93e1ddcd62500cba0991bc9a17f6718dbcec7bc52b714d4703464cb4f9539bf2"} Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.776499 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"deb9f370792598608cf2e0c0fe9975eaf8b95dc94b12eef8c55112df4aa9c998"} Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.776528 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"2e2b783d14d219076ec2f389ad2ff4e6d42a1c7ac369ce3dc23db350f24e7ede"} Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.778902 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.781515 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7079548fe9529583a1c129f57838c24cf06d1cb55a97c05c63a05d0cc2e7ec9b"} Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.781967 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.782929 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"505e78ddf4f09331c2388f1fa7083f24e2dc672f77671a10eec53d6ba4bb1bef"} Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.800240 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:04Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.814576 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:04Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.829570 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:04Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.840598 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee17b4a9a8999e5f76ceaf69dbd13acf80ed0e95e72f3e1428ce38c5d5982788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://382256e0fa1a4c26610227b9ae96288ccbfbbd97aa3cbd5fd4f18cb4228a234b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:04Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.852272 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d40160e-075f-44b7-9d9d-3291e7774f27\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7762237003a5a256ff20a16a4869462e2e0dd1426aa391f7ec74ed9b1f572d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7762237003a5a256ff20a16a4869462e2e0dd1426aa391f7ec74ed9b1f572d4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1129 04:39:02.908489 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1129 04:39:02.908637 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 04:39:02.909677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703886220/tls.crt::/tmp/serving-cert-1703886220/tls.key\\\\\\\"\\\\nI1129 04:39:03.077876 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 04:39:03.080368 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 04:39:03.080390 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 04:39:03.080424 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 04:39:03.080432 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 04:39:03.086120 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 04:39:03.086240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086412 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 04:39:03.086502 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 04:39:03.086563 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 04:39:03.086622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1129 04:39:03.086141 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1129 04:39:03.086836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:38:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:04Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.863133 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:04Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.872466 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:04Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.884503 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d40160e-075f-44b7-9d9d-3291e7774f27\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7079548fe9529583a1c129f57838c24cf06d1cb55a97c05c63a05d0cc2e7ec9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7762237003a5a256ff20a16a4869462e2e0dd1426aa391f7ec74ed9b1f572d4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1129 04:39:02.908489 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1129 04:39:02.908637 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 04:39:02.909677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703886220/tls.crt::/tmp/serving-cert-1703886220/tls.key\\\\\\\"\\\\nI1129 04:39:03.077876 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 04:39:03.080368 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 04:39:03.080390 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 04:39:03.080424 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 04:39:03.080432 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 04:39:03.086120 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 04:39:03.086240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086412 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 04:39:03.086502 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 04:39:03.086563 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 04:39:03.086622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1129 04:39:03.086141 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1129 04:39:03.086836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:38:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:04Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.895970 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:04Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.908131 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:04Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.919979 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://deb9f370792598608cf2e0c0fe9975eaf8b95dc94b12eef8c55112df4aa9c998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:04Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.932433 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:04Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.967079 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:04Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:04 crc kubenswrapper[4799]: I1129 04:39:04.980551 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee17b4a9a8999e5f76ceaf69dbd13acf80ed0e95e72f3e1428ce38c5d5982788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://382256e0fa1a4c26610227b9ae96288ccbfbbd97aa3cbd5fd4f18cb4228a234b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:04Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:05 crc kubenswrapper[4799]: I1129 04:39:05.426727 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 04:39:05 crc kubenswrapper[4799]: I1129 04:39:05.426849 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 04:39:05 crc kubenswrapper[4799]: I1129 04:39:05.426883 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:39:05 crc kubenswrapper[4799]: I1129 04:39:05.426912 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:39:05 crc kubenswrapper[4799]: I1129 04:39:05.426940 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 04:39:05 crc kubenswrapper[4799]: E1129 04:39:05.427116 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 29 04:39:05 crc kubenswrapper[4799]: E1129 04:39:05.427145 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 29 04:39:05 crc kubenswrapper[4799]: E1129 04:39:05.427160 4799 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 04:39:05 crc kubenswrapper[4799]: E1129 04:39:05.427268 4799 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 29 04:39:05 crc kubenswrapper[4799]: E1129 04:39:05.427355 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-29 04:39:07.427330471 +0000 UTC m=+23.070260871 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 29 04:39:05 crc kubenswrapper[4799]: E1129 04:39:05.427366 4799 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 29 04:39:05 crc kubenswrapper[4799]: E1129 04:39:05.427492 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-29 04:39:07.427458734 +0000 UTC m=+23.070389344 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 29 04:39:05 crc kubenswrapper[4799]: E1129 04:39:05.427636 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 29 04:39:05 crc kubenswrapper[4799]: E1129 04:39:05.427723 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 29 04:39:05 crc kubenswrapper[4799]: E1129 04:39:05.427822 4799 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 04:39:05 crc kubenswrapper[4799]: E1129 04:39:05.427637 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 04:39:07.427622458 +0000 UTC m=+23.070553098 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:39:05 crc kubenswrapper[4799]: E1129 04:39:05.428052 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-29 04:39:07.428038038 +0000 UTC m=+23.070968628 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 04:39:05 crc kubenswrapper[4799]: E1129 04:39:05.428161 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-29 04:39:07.428147041 +0000 UTC m=+23.071077641 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 04:39:05 crc kubenswrapper[4799]: I1129 04:39:05.657844 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:39:05 crc kubenswrapper[4799]: I1129 04:39:05.657844 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 04:39:05 crc kubenswrapper[4799]: E1129 04:39:05.658047 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 04:39:05 crc kubenswrapper[4799]: E1129 04:39:05.657951 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 04:39:05 crc kubenswrapper[4799]: I1129 04:39:05.657844 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 04:39:05 crc kubenswrapper[4799]: E1129 04:39:05.658142 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 04:39:06 crc kubenswrapper[4799]: I1129 04:39:06.791267 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"396c1c8cbd638c7dc4d8eebded6a17449d9abdc049161578581f1d8c3e31d9a6"} Nov 29 04:39:06 crc kubenswrapper[4799]: I1129 04:39:06.816306 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:06Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:06 crc kubenswrapper[4799]: I1129 04:39:06.829594 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee17b4a9a8999e5f76ceaf69dbd13acf80ed0e95e72f3e1428ce38c5d5982788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://382256e0fa1a4c26610227b9ae96288ccbfbbd97aa3cbd5fd4f18cb4228a234b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:06Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:06 crc kubenswrapper[4799]: I1129 04:39:06.842637 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://deb9f370792598608cf2e0c0fe9975eaf8b95dc94b12eef8c55112df4aa9c998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:06Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:06 crc kubenswrapper[4799]: I1129 04:39:06.863780 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:06Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:06 crc kubenswrapper[4799]: I1129 04:39:06.877623 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:06Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:06 crc kubenswrapper[4799]: I1129 04:39:06.895055 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d40160e-075f-44b7-9d9d-3291e7774f27\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7079548fe9529583a1c129f57838c24cf06d1cb55a97c05c63a05d0cc2e7ec9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7762237003a5a256ff20a16a4869462e2e0dd1426aa391f7ec74ed9b1f572d4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1129 04:39:02.908489 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1129 04:39:02.908637 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 04:39:02.909677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703886220/tls.crt::/tmp/serving-cert-1703886220/tls.key\\\\\\\"\\\\nI1129 04:39:03.077876 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 04:39:03.080368 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 04:39:03.080390 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 04:39:03.080424 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 04:39:03.080432 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 04:39:03.086120 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 04:39:03.086240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086412 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 04:39:03.086502 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 04:39:03.086563 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 04:39:03.086622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1129 04:39:03.086141 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1129 04:39:03.086836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:38:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:06Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:06 crc kubenswrapper[4799]: I1129 04:39:06.909770 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://396c1c8cbd638c7dc4d8eebded6a17449d9abdc049161578581f1d8c3e31d9a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:06Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.445015 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.445103 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.445141 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 04:39:07 crc kubenswrapper[4799]: E1129 04:39:07.445172 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 04:39:11.445151392 +0000 UTC m=+27.088081792 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.445202 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.445236 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:39:07 crc kubenswrapper[4799]: E1129 04:39:07.445254 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 29 04:39:07 crc kubenswrapper[4799]: E1129 04:39:07.445254 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 29 04:39:07 crc kubenswrapper[4799]: E1129 04:39:07.445285 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 29 04:39:07 crc kubenswrapper[4799]: E1129 04:39:07.445287 4799 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 29 04:39:07 crc kubenswrapper[4799]: E1129 04:39:07.445306 4799 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 04:39:07 crc kubenswrapper[4799]: E1129 04:39:07.445269 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 29 04:39:07 crc kubenswrapper[4799]: E1129 04:39:07.445434 4799 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 04:39:07 crc kubenswrapper[4799]: E1129 04:39:07.445319 4799 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 29 04:39:07 crc kubenswrapper[4799]: E1129 04:39:07.445336 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-29 04:39:11.445326766 +0000 UTC m=+27.088257166 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 29 04:39:07 crc kubenswrapper[4799]: E1129 04:39:07.445568 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-29 04:39:11.445535571 +0000 UTC m=+27.088465971 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 04:39:07 crc kubenswrapper[4799]: E1129 04:39:07.445595 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-29 04:39:11.445586003 +0000 UTC m=+27.088516403 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 04:39:07 crc kubenswrapper[4799]: E1129 04:39:07.445620 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-29 04:39:11.445611003 +0000 UTC m=+27.088541403 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.496211 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.498065 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.498124 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.498137 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.498207 4799 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.505611 4799 kubelet_node_status.go:115] "Node was previously registered" node="crc" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.505882 4799 kubelet_node_status.go:79] "Successfully registered node" node="crc" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.506808 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.506843 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.506854 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.506870 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.506881 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:07Z","lastTransitionTime":"2025-11-29T04:39:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:07 crc kubenswrapper[4799]: E1129 04:39:07.526807 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6f4e4672-da32-440c-8c4f-6e7f4942f55e\\\",\\\"systemUUID\\\":\\\"882c4d8e-57b7-4312-8499-40d1f85d56c9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:07Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.533125 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.533173 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.533182 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.533196 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.533207 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:07Z","lastTransitionTime":"2025-11-29T04:39:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:07 crc kubenswrapper[4799]: E1129 04:39:07.548049 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6f4e4672-da32-440c-8c4f-6e7f4942f55e\\\",\\\"systemUUID\\\":\\\"882c4d8e-57b7-4312-8499-40d1f85d56c9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:07Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.550907 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.550940 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.550952 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.550964 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.550974 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:07Z","lastTransitionTime":"2025-11-29T04:39:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:07 crc kubenswrapper[4799]: E1129 04:39:07.568073 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6f4e4672-da32-440c-8c4f-6e7f4942f55e\\\",\\\"systemUUID\\\":\\\"882c4d8e-57b7-4312-8499-40d1f85d56c9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:07Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.571874 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.571918 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.571933 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.571953 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.571969 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:07Z","lastTransitionTime":"2025-11-29T04:39:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:07 crc kubenswrapper[4799]: E1129 04:39:07.585295 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6f4e4672-da32-440c-8c4f-6e7f4942f55e\\\",\\\"systemUUID\\\":\\\"882c4d8e-57b7-4312-8499-40d1f85d56c9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:07Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.588477 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.588498 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.588506 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.588517 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.588526 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:07Z","lastTransitionTime":"2025-11-29T04:39:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:07 crc kubenswrapper[4799]: E1129 04:39:07.598873 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6f4e4672-da32-440c-8c4f-6e7f4942f55e\\\",\\\"systemUUID\\\":\\\"882c4d8e-57b7-4312-8499-40d1f85d56c9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:07Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:07 crc kubenswrapper[4799]: E1129 04:39:07.598978 4799 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.600157 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.600204 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.600224 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.600242 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.600258 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:07Z","lastTransitionTime":"2025-11-29T04:39:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.657971 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.657971 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.658054 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 04:39:07 crc kubenswrapper[4799]: E1129 04:39:07.658134 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 04:39:07 crc kubenswrapper[4799]: E1129 04:39:07.658211 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 04:39:07 crc kubenswrapper[4799]: E1129 04:39:07.658286 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.702758 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.702819 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.702831 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.702847 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.702858 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:07Z","lastTransitionTime":"2025-11-29T04:39:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.805987 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.806025 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.806033 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.806047 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.806056 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:07Z","lastTransitionTime":"2025-11-29T04:39:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.908530 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.908578 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.908592 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.908610 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:07 crc kubenswrapper[4799]: I1129 04:39:07.908623 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:07Z","lastTransitionTime":"2025-11-29T04:39:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.010961 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.011001 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.011010 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.011025 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.011034 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:08Z","lastTransitionTime":"2025-11-29T04:39:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.113763 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.113913 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.113941 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.113972 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.113994 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:08Z","lastTransitionTime":"2025-11-29T04:39:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.215816 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.215862 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.215877 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.215898 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.215913 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:08Z","lastTransitionTime":"2025-11-29T04:39:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.318991 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.319032 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.319042 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.319085 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.319099 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:08Z","lastTransitionTime":"2025-11-29T04:39:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.421938 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.422005 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.422025 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.422053 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.422070 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:08Z","lastTransitionTime":"2025-11-29T04:39:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.524835 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.524878 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.524888 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.524902 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.524911 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:08Z","lastTransitionTime":"2025-11-29T04:39:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.627229 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.627259 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.627266 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.627280 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.627288 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:08Z","lastTransitionTime":"2025-11-29T04:39:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.729893 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.729935 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.729943 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.729957 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.729967 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:08Z","lastTransitionTime":"2025-11-29T04:39:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.832376 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.832425 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.832434 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.832446 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.832454 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:08Z","lastTransitionTime":"2025-11-29T04:39:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.935187 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.935227 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.935235 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.935247 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.935256 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:08Z","lastTransitionTime":"2025-11-29T04:39:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.967930 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.973026 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.976271 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.985486 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d40160e-075f-44b7-9d9d-3291e7774f27\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7079548fe9529583a1c129f57838c24cf06d1cb55a97c05c63a05d0cc2e7ec9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7762237003a5a256ff20a16a4869462e2e0dd1426aa391f7ec74ed9b1f572d4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1129 04:39:02.908489 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1129 04:39:02.908637 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 04:39:02.909677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703886220/tls.crt::/tmp/serving-cert-1703886220/tls.key\\\\\\\"\\\\nI1129 04:39:03.077876 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 04:39:03.080368 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 04:39:03.080390 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 04:39:03.080424 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 04:39:03.080432 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 04:39:03.086120 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 04:39:03.086240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086412 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 04:39:03.086502 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 04:39:03.086563 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 04:39:03.086622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1129 04:39:03.086141 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1129 04:39:03.086836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:38:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:08Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:08 crc kubenswrapper[4799]: I1129 04:39:08.997394 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://396c1c8cbd638c7dc4d8eebded6a17449d9abdc049161578581f1d8c3e31d9a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:08Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.011438 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:09Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.023170 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://deb9f370792598608cf2e0c0fe9975eaf8b95dc94b12eef8c55112df4aa9c998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:09Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.037190 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:09Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.038422 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.038467 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.038479 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.038495 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.038508 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:09Z","lastTransitionTime":"2025-11-29T04:39:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.050854 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:09Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.065482 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee17b4a9a8999e5f76ceaf69dbd13acf80ed0e95e72f3e1428ce38c5d5982788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://382256e0fa1a4c26610227b9ae96288ccbfbbd97aa3cbd5fd4f18cb4228a234b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:09Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.078570 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://396c1c8cbd638c7dc4d8eebded6a17449d9abdc049161578581f1d8c3e31d9a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:09Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.094918 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:09Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.108629 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d40160e-075f-44b7-9d9d-3291e7774f27\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7079548fe9529583a1c129f57838c24cf06d1cb55a97c05c63a05d0cc2e7ec9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7762237003a5a256ff20a16a4869462e2e0dd1426aa391f7ec74ed9b1f572d4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1129 04:39:02.908489 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1129 04:39:02.908637 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 04:39:02.909677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703886220/tls.crt::/tmp/serving-cert-1703886220/tls.key\\\\\\\"\\\\nI1129 04:39:03.077876 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 04:39:03.080368 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 04:39:03.080390 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 04:39:03.080424 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 04:39:03.080432 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 04:39:03.086120 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 04:39:03.086240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086412 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 04:39:03.086502 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 04:39:03.086563 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 04:39:03.086622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1129 04:39:03.086141 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1129 04:39:03.086836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:38:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:09Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.120095 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22ed28a-3748-4eb9-b3a7-c02aa38f7e94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33c0540ccb097b15f65369ff537aef4ae621cc828ba314eb61f0a0037608cadf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c241d29297a73dce12bd0bc2863c56475eb35b8fe8c93991ba0ef8695e38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f42a55d797c6c94af3a5eac1bf08aa38a82359b7ddd0ba32dcc797218ceee6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebed9ec415bb433935da36df8014f9087236cea9788b2048cbff8714f8ad0df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:09Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.132450 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://deb9f370792598608cf2e0c0fe9975eaf8b95dc94b12eef8c55112df4aa9c998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:09Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.140046 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.140082 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.140092 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.140106 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.140119 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:09Z","lastTransitionTime":"2025-11-29T04:39:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.145302 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:09Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.157231 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:09Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.174728 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee17b4a9a8999e5f76ceaf69dbd13acf80ed0e95e72f3e1428ce38c5d5982788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://382256e0fa1a4c26610227b9ae96288ccbfbbd97aa3cbd5fd4f18cb4228a234b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:09Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.242278 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.242332 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.242341 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.242352 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.242361 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:09Z","lastTransitionTime":"2025-11-29T04:39:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.344388 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.344430 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.344439 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.344469 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.344484 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:09Z","lastTransitionTime":"2025-11-29T04:39:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.399911 4799 csr.go:261] certificate signing request csr-drrcf is approved, waiting to be issued Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.416171 4799 csr.go:257] certificate signing request csr-drrcf is issued Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.446930 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.446974 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.446984 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.447001 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.447012 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:09Z","lastTransitionTime":"2025-11-29T04:39:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.549932 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.549982 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.549995 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.550014 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.550031 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:09Z","lastTransitionTime":"2025-11-29T04:39:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.652283 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.652330 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.652346 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.652367 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.652379 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:09Z","lastTransitionTime":"2025-11-29T04:39:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.658506 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.658548 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.658518 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:39:09 crc kubenswrapper[4799]: E1129 04:39:09.658630 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 04:39:09 crc kubenswrapper[4799]: E1129 04:39:09.658681 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 04:39:09 crc kubenswrapper[4799]: E1129 04:39:09.658754 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.754368 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.754427 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.754442 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.754457 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.754469 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:09Z","lastTransitionTime":"2025-11-29T04:39:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.805266 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-qt2dl"] Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.805513 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-v7lb4"] Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.805692 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.805730 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-v7lb4" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.808458 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.808563 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.808669 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.809444 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.809712 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.810131 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.810278 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.810876 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.828760 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt2dl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6c9236-5f7d-46c5-b080-4667620d874e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt2dl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:09Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.846862 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://deb9f370792598608cf2e0c0fe9975eaf8b95dc94b12eef8c55112df4aa9c998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:09Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.856702 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.856741 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.856752 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.856769 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.856779 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:09Z","lastTransitionTime":"2025-11-29T04:39:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.861008 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:09Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.865328 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/55156232-4c80-45a7-aad5-a5fa2793e4ed-hosts-file\") pod \"node-resolver-v7lb4\" (UID: \"55156232-4c80-45a7-aad5-a5fa2793e4ed\") " pod="openshift-dns/node-resolver-v7lb4" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.865373 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3b6c9236-5f7d-46c5-b080-4667620d874e-etc-kubernetes\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.865415 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/3b6c9236-5f7d-46c5-b080-4667620d874e-os-release\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.865449 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/3b6c9236-5f7d-46c5-b080-4667620d874e-multus-socket-dir-parent\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.865468 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/3b6c9236-5f7d-46c5-b080-4667620d874e-host-run-k8s-cni-cncf-io\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.865487 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/3b6c9236-5f7d-46c5-b080-4667620d874e-host-var-lib-cni-multus\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.865515 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/3b6c9236-5f7d-46c5-b080-4667620d874e-hostroot\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.865536 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/3b6c9236-5f7d-46c5-b080-4667620d874e-multus-conf-dir\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.865560 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/3b6c9236-5f7d-46c5-b080-4667620d874e-host-var-lib-kubelet\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.865670 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3b6c9236-5f7d-46c5-b080-4667620d874e-host-var-lib-cni-bin\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.865766 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2dtw\" (UniqueName: \"kubernetes.io/projected/55156232-4c80-45a7-aad5-a5fa2793e4ed-kube-api-access-d2dtw\") pod \"node-resolver-v7lb4\" (UID: \"55156232-4c80-45a7-aad5-a5fa2793e4ed\") " pod="openshift-dns/node-resolver-v7lb4" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.865810 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvz6r\" (UniqueName: \"kubernetes.io/projected/3b6c9236-5f7d-46c5-b080-4667620d874e-kube-api-access-xvz6r\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.865919 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/3b6c9236-5f7d-46c5-b080-4667620d874e-multus-daemon-config\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.865982 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/3b6c9236-5f7d-46c5-b080-4667620d874e-host-run-multus-certs\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.866042 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/3b6c9236-5f7d-46c5-b080-4667620d874e-system-cni-dir\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.866063 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3b6c9236-5f7d-46c5-b080-4667620d874e-host-run-netns\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.866084 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/3b6c9236-5f7d-46c5-b080-4667620d874e-cnibin\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.866103 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/3b6c9236-5f7d-46c5-b080-4667620d874e-cni-binary-copy\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.866134 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/3b6c9236-5f7d-46c5-b080-4667620d874e-multus-cni-dir\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.875402 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:09Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.886830 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee17b4a9a8999e5f76ceaf69dbd13acf80ed0e95e72f3e1428ce38c5d5982788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://382256e0fa1a4c26610227b9ae96288ccbfbbd97aa3cbd5fd4f18cb4228a234b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:09Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.901102 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d40160e-075f-44b7-9d9d-3291e7774f27\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7079548fe9529583a1c129f57838c24cf06d1cb55a97c05c63a05d0cc2e7ec9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7762237003a5a256ff20a16a4869462e2e0dd1426aa391f7ec74ed9b1f572d4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1129 04:39:02.908489 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1129 04:39:02.908637 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 04:39:02.909677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703886220/tls.crt::/tmp/serving-cert-1703886220/tls.key\\\\\\\"\\\\nI1129 04:39:03.077876 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 04:39:03.080368 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 04:39:03.080390 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 04:39:03.080424 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 04:39:03.080432 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 04:39:03.086120 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 04:39:03.086240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086412 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 04:39:03.086502 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 04:39:03.086563 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 04:39:03.086622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1129 04:39:03.086141 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1129 04:39:03.086836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:38:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:09Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.918661 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22ed28a-3748-4eb9-b3a7-c02aa38f7e94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33c0540ccb097b15f65369ff537aef4ae621cc828ba314eb61f0a0037608cadf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c241d29297a73dce12bd0bc2863c56475eb35b8fe8c93991ba0ef8695e38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f42a55d797c6c94af3a5eac1bf08aa38a82359b7ddd0ba32dcc797218ceee6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebed9ec415bb433935da36df8014f9087236cea9788b2048cbff8714f8ad0df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:09Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.933506 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://396c1c8cbd638c7dc4d8eebded6a17449d9abdc049161578581f1d8c3e31d9a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:09Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.949596 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:09Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.958394 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.958445 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.958457 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.958474 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.958487 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:09Z","lastTransitionTime":"2025-11-29T04:39:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.964363 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7lb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55156232-4c80-45a7-aad5-a5fa2793e4ed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d2dtw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7lb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:09Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.967189 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/3b6c9236-5f7d-46c5-b080-4667620d874e-os-release\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.967233 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/3b6c9236-5f7d-46c5-b080-4667620d874e-multus-socket-dir-parent\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.967255 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/3b6c9236-5f7d-46c5-b080-4667620d874e-host-run-k8s-cni-cncf-io\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.967274 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/3b6c9236-5f7d-46c5-b080-4667620d874e-host-var-lib-cni-multus\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.967295 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/3b6c9236-5f7d-46c5-b080-4667620d874e-hostroot\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.967310 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/3b6c9236-5f7d-46c5-b080-4667620d874e-multus-conf-dir\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.967329 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/3b6c9236-5f7d-46c5-b080-4667620d874e-host-var-lib-kubelet\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.967337 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/3b6c9236-5f7d-46c5-b080-4667620d874e-os-release\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.967346 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3b6c9236-5f7d-46c5-b080-4667620d874e-host-var-lib-cni-bin\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.967370 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3b6c9236-5f7d-46c5-b080-4667620d874e-host-var-lib-cni-bin\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.967411 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/3b6c9236-5f7d-46c5-b080-4667620d874e-host-var-lib-kubelet\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.967429 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/3b6c9236-5f7d-46c5-b080-4667620d874e-host-run-k8s-cni-cncf-io\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.967445 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/3b6c9236-5f7d-46c5-b080-4667620d874e-host-var-lib-cni-multus\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.967441 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/3b6c9236-5f7d-46c5-b080-4667620d874e-hostroot\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.967470 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/3b6c9236-5f7d-46c5-b080-4667620d874e-multus-conf-dir\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.967535 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2dtw\" (UniqueName: \"kubernetes.io/projected/55156232-4c80-45a7-aad5-a5fa2793e4ed-kube-api-access-d2dtw\") pod \"node-resolver-v7lb4\" (UID: \"55156232-4c80-45a7-aad5-a5fa2793e4ed\") " pod="openshift-dns/node-resolver-v7lb4" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.967583 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvz6r\" (UniqueName: \"kubernetes.io/projected/3b6c9236-5f7d-46c5-b080-4667620d874e-kube-api-access-xvz6r\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.967464 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/3b6c9236-5f7d-46c5-b080-4667620d874e-multus-socket-dir-parent\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.967600 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/3b6c9236-5f7d-46c5-b080-4667620d874e-multus-daemon-config\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.967700 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/3b6c9236-5f7d-46c5-b080-4667620d874e-host-run-multus-certs\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.967750 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/3b6c9236-5f7d-46c5-b080-4667620d874e-system-cni-dir\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.967766 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/3b6c9236-5f7d-46c5-b080-4667620d874e-host-run-multus-certs\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.967781 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3b6c9236-5f7d-46c5-b080-4667620d874e-host-run-netns\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.967818 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/3b6c9236-5f7d-46c5-b080-4667620d874e-system-cni-dir\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.967826 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/3b6c9236-5f7d-46c5-b080-4667620d874e-cnibin\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.967890 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/3b6c9236-5f7d-46c5-b080-4667620d874e-cni-binary-copy\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.967912 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/3b6c9236-5f7d-46c5-b080-4667620d874e-cnibin\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.967916 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/3b6c9236-5f7d-46c5-b080-4667620d874e-multus-cni-dir\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.967951 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/55156232-4c80-45a7-aad5-a5fa2793e4ed-hosts-file\") pod \"node-resolver-v7lb4\" (UID: \"55156232-4c80-45a7-aad5-a5fa2793e4ed\") " pod="openshift-dns/node-resolver-v7lb4" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.967975 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3b6c9236-5f7d-46c5-b080-4667620d874e-etc-kubernetes\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.967903 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3b6c9236-5f7d-46c5-b080-4667620d874e-host-run-netns\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.968009 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/55156232-4c80-45a7-aad5-a5fa2793e4ed-hosts-file\") pod \"node-resolver-v7lb4\" (UID: \"55156232-4c80-45a7-aad5-a5fa2793e4ed\") " pod="openshift-dns/node-resolver-v7lb4" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.968042 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3b6c9236-5f7d-46c5-b080-4667620d874e-etc-kubernetes\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.968144 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/3b6c9236-5f7d-46c5-b080-4667620d874e-multus-cni-dir\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.968437 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/3b6c9236-5f7d-46c5-b080-4667620d874e-multus-daemon-config\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.968776 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/3b6c9236-5f7d-46c5-b080-4667620d874e-cni-binary-copy\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.978599 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee17b4a9a8999e5f76ceaf69dbd13acf80ed0e95e72f3e1428ce38c5d5982788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://382256e0fa1a4c26610227b9ae96288ccbfbbd97aa3cbd5fd4f18cb4228a234b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:09Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.995464 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2dtw\" (UniqueName: \"kubernetes.io/projected/55156232-4c80-45a7-aad5-a5fa2793e4ed-kube-api-access-d2dtw\") pod \"node-resolver-v7lb4\" (UID: \"55156232-4c80-45a7-aad5-a5fa2793e4ed\") " pod="openshift-dns/node-resolver-v7lb4" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.997767 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt2dl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6c9236-5f7d-46c5-b080-4667620d874e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt2dl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:09Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:09 crc kubenswrapper[4799]: I1129 04:39:09.999368 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvz6r\" (UniqueName: \"kubernetes.io/projected/3b6c9236-5f7d-46c5-b080-4667620d874e-kube-api-access-xvz6r\") pod \"multus-qt2dl\" (UID: \"3b6c9236-5f7d-46c5-b080-4667620d874e\") " pod="openshift-multus/multus-qt2dl" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.009723 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://deb9f370792598608cf2e0c0fe9975eaf8b95dc94b12eef8c55112df4aa9c998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:10Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.039550 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:10Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.060741 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.060796 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.060809 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.060826 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.060838 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:10Z","lastTransitionTime":"2025-11-29T04:39:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.077587 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:10Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.116545 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7lb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55156232-4c80-45a7-aad5-a5fa2793e4ed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d2dtw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7lb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:10Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.119667 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-qt2dl" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.124755 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-v7lb4" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.165338 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.165520 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.165583 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.165661 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.165735 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:10Z","lastTransitionTime":"2025-11-29T04:39:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.191518 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d40160e-075f-44b7-9d9d-3291e7774f27\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7079548fe9529583a1c129f57838c24cf06d1cb55a97c05c63a05d0cc2e7ec9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7762237003a5a256ff20a16a4869462e2e0dd1426aa391f7ec74ed9b1f572d4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1129 04:39:02.908489 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1129 04:39:02.908637 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 04:39:02.909677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703886220/tls.crt::/tmp/serving-cert-1703886220/tls.key\\\\\\\"\\\\nI1129 04:39:03.077876 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 04:39:03.080368 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 04:39:03.080390 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 04:39:03.080424 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 04:39:03.080432 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 04:39:03.086120 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 04:39:03.086240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086412 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 04:39:03.086502 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 04:39:03.086563 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 04:39:03.086622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1129 04:39:03.086141 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1129 04:39:03.086836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:38:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:10Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.211930 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-s5d7s"] Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.212541 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-s5d7s" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.214190 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-fht8l"] Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.214894 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.227517 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-26t88"] Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.228279 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.228348 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.228282 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.228283 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.228511 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.229608 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.230120 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.230530 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-26t88" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.230958 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.235742 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.235966 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.236096 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.240825 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.240907 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22ed28a-3748-4eb9-b3a7-c02aa38f7e94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33c0540ccb097b15f65369ff537aef4ae621cc828ba314eb61f0a0037608cadf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c241d29297a73dce12bd0bc2863c56475eb35b8fe8c93991ba0ef8695e38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f42a55d797c6c94af3a5eac1bf08aa38a82359b7ddd0ba32dcc797218ceee6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebed9ec415bb433935da36df8014f9087236cea9788b2048cbff8714f8ad0df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:10Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.240967 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.240985 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.259267 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://396c1c8cbd638c7dc4d8eebded6a17449d9abdc049161578581f1d8c3e31d9a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:10Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.268198 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.268226 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.268234 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.268248 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.268257 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:10Z","lastTransitionTime":"2025-11-29T04:39:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.270668 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:10Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.270758 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/b333a2e0-0da7-4d88-9539-0b4cc12bc2e8-rootfs\") pod \"machine-config-daemon-26t88\" (UID: \"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8\") " pod="openshift-machine-config-operator/machine-config-daemon-26t88" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.270780 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-host-run-netns\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.270811 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/0af3d308-36b9-4622-8544-635f7b24d687-cni-binary-copy\") pod \"multus-additional-cni-plugins-s5d7s\" (UID: \"0af3d308-36b9-4622-8544-635f7b24d687\") " pod="openshift-multus/multus-additional-cni-plugins-s5d7s" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.270831 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-log-socket\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.270846 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/0af3d308-36b9-4622-8544-635f7b24d687-tuning-conf-dir\") pod \"multus-additional-cni-plugins-s5d7s\" (UID: \"0af3d308-36b9-4622-8544-635f7b24d687\") " pod="openshift-multus/multus-additional-cni-plugins-s5d7s" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.270860 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/0af3d308-36b9-4622-8544-635f7b24d687-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-s5d7s\" (UID: \"0af3d308-36b9-4622-8544-635f7b24d687\") " pod="openshift-multus/multus-additional-cni-plugins-s5d7s" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.270888 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-host-cni-netd\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.270919 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b333a2e0-0da7-4d88-9539-0b4cc12bc2e8-proxy-tls\") pod \"machine-config-daemon-26t88\" (UID: \"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8\") " pod="openshift-machine-config-operator/machine-config-daemon-26t88" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.270936 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-run-ovn\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.270961 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-run-openvswitch\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.270977 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-host-run-ovn-kubernetes\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.270998 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xlq24\" (UniqueName: \"kubernetes.io/projected/0af3d308-36b9-4622-8544-635f7b24d687-kube-api-access-xlq24\") pod \"multus-additional-cni-plugins-s5d7s\" (UID: \"0af3d308-36b9-4622-8544-635f7b24d687\") " pod="openshift-multus/multus-additional-cni-plugins-s5d7s" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.271015 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-var-lib-openvswitch\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.271046 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-systemd-units\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.271075 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.271094 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bd007c61-4689-4179-95ad-44a9724e8e7e-ovn-node-metrics-cert\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.271119 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6v22j\" (UniqueName: \"kubernetes.io/projected/b333a2e0-0da7-4d88-9539-0b4cc12bc2e8-kube-api-access-6v22j\") pod \"machine-config-daemon-26t88\" (UID: \"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8\") " pod="openshift-machine-config-operator/machine-config-daemon-26t88" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.271136 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bd007c61-4689-4179-95ad-44a9724e8e7e-env-overrides\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.271170 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-host-cni-bin\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.271188 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-node-log\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.271264 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-host-slash\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.271299 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-etc-openvswitch\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.271333 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-host-kubelet\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.271348 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0af3d308-36b9-4622-8544-635f7b24d687-system-cni-dir\") pod \"multus-additional-cni-plugins-s5d7s\" (UID: \"0af3d308-36b9-4622-8544-635f7b24d687\") " pod="openshift-multus/multus-additional-cni-plugins-s5d7s" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.271366 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/0af3d308-36b9-4622-8544-635f7b24d687-cnibin\") pod \"multus-additional-cni-plugins-s5d7s\" (UID: \"0af3d308-36b9-4622-8544-635f7b24d687\") " pod="openshift-multus/multus-additional-cni-plugins-s5d7s" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.271381 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b333a2e0-0da7-4d88-9539-0b4cc12bc2e8-mcd-auth-proxy-config\") pod \"machine-config-daemon-26t88\" (UID: \"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8\") " pod="openshift-machine-config-operator/machine-config-daemon-26t88" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.271402 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/bd007c61-4689-4179-95ad-44a9724e8e7e-ovnkube-script-lib\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.271418 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpd4j\" (UniqueName: \"kubernetes.io/projected/bd007c61-4689-4179-95ad-44a9724e8e7e-kube-api-access-jpd4j\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.271432 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/0af3d308-36b9-4622-8544-635f7b24d687-os-release\") pod \"multus-additional-cni-plugins-s5d7s\" (UID: \"0af3d308-36b9-4622-8544-635f7b24d687\") " pod="openshift-multus/multus-additional-cni-plugins-s5d7s" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.271456 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-run-systemd\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.271468 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bd007c61-4689-4179-95ad-44a9724e8e7e-ovnkube-config\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.280857 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://396c1c8cbd638c7dc4d8eebded6a17449d9abdc049161578581f1d8c3e31d9a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:10Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.293178 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d40160e-075f-44b7-9d9d-3291e7774f27\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7079548fe9529583a1c129f57838c24cf06d1cb55a97c05c63a05d0cc2e7ec9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7762237003a5a256ff20a16a4869462e2e0dd1426aa391f7ec74ed9b1f572d4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1129 04:39:02.908489 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1129 04:39:02.908637 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 04:39:02.909677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703886220/tls.crt::/tmp/serving-cert-1703886220/tls.key\\\\\\\"\\\\nI1129 04:39:03.077876 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 04:39:03.080368 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 04:39:03.080390 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 04:39:03.080424 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 04:39:03.080432 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 04:39:03.086120 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 04:39:03.086240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086412 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 04:39:03.086502 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 04:39:03.086563 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 04:39:03.086622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1129 04:39:03.086141 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1129 04:39:03.086836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:38:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:10Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.305382 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:10Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.322338 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd007c61-4689-4179-95ad-44a9724e8e7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fht8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:10Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.333092 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:10Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.341676 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7lb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55156232-4c80-45a7-aad5-a5fa2793e4ed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d2dtw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7lb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:10Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.355064 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-26t88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:10Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.370924 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.370958 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.370967 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.370982 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.370991 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:10Z","lastTransitionTime":"2025-11-29T04:39:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.372180 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-run-systemd\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.372212 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bd007c61-4689-4179-95ad-44a9724e8e7e-ovnkube-config\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.372227 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/bd007c61-4689-4179-95ad-44a9724e8e7e-ovnkube-script-lib\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.372245 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpd4j\" (UniqueName: \"kubernetes.io/projected/bd007c61-4689-4179-95ad-44a9724e8e7e-kube-api-access-jpd4j\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.372259 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/0af3d308-36b9-4622-8544-635f7b24d687-os-release\") pod \"multus-additional-cni-plugins-s5d7s\" (UID: \"0af3d308-36b9-4622-8544-635f7b24d687\") " pod="openshift-multus/multus-additional-cni-plugins-s5d7s" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.372283 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/b333a2e0-0da7-4d88-9539-0b4cc12bc2e8-rootfs\") pod \"machine-config-daemon-26t88\" (UID: \"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8\") " pod="openshift-machine-config-operator/machine-config-daemon-26t88" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.372275 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-run-systemd\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.372308 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/0af3d308-36b9-4622-8544-635f7b24d687-cni-binary-copy\") pod \"multus-additional-cni-plugins-s5d7s\" (UID: \"0af3d308-36b9-4622-8544-635f7b24d687\") " pod="openshift-multus/multus-additional-cni-plugins-s5d7s" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.372367 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-host-run-netns\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.372390 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-log-socket\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.372414 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/0af3d308-36b9-4622-8544-635f7b24d687-tuning-conf-dir\") pod \"multus-additional-cni-plugins-s5d7s\" (UID: \"0af3d308-36b9-4622-8544-635f7b24d687\") " pod="openshift-multus/multus-additional-cni-plugins-s5d7s" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.372436 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/0af3d308-36b9-4622-8544-635f7b24d687-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-s5d7s\" (UID: \"0af3d308-36b9-4622-8544-635f7b24d687\") " pod="openshift-multus/multus-additional-cni-plugins-s5d7s" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.372460 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b333a2e0-0da7-4d88-9539-0b4cc12bc2e8-proxy-tls\") pod \"machine-config-daemon-26t88\" (UID: \"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8\") " pod="openshift-machine-config-operator/machine-config-daemon-26t88" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.372479 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-run-ovn\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.372503 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-host-cni-netd\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.372526 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-run-openvswitch\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.372544 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-host-run-ovn-kubernetes\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.372560 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xlq24\" (UniqueName: \"kubernetes.io/projected/0af3d308-36b9-4622-8544-635f7b24d687-kube-api-access-xlq24\") pod \"multus-additional-cni-plugins-s5d7s\" (UID: \"0af3d308-36b9-4622-8544-635f7b24d687\") " pod="openshift-multus/multus-additional-cni-plugins-s5d7s" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.372578 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-systemd-units\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.372594 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-var-lib-openvswitch\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.372610 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.372626 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bd007c61-4689-4179-95ad-44a9724e8e7e-ovn-node-metrics-cert\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.372655 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6v22j\" (UniqueName: \"kubernetes.io/projected/b333a2e0-0da7-4d88-9539-0b4cc12bc2e8-kube-api-access-6v22j\") pod \"machine-config-daemon-26t88\" (UID: \"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8\") " pod="openshift-machine-config-operator/machine-config-daemon-26t88" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.372679 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-host-cni-bin\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.372721 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bd007c61-4689-4179-95ad-44a9724e8e7e-env-overrides\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.372737 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-node-log\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.372765 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-host-slash\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.372780 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-etc-openvswitch\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.372825 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-host-kubelet\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.372845 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0af3d308-36b9-4622-8544-635f7b24d687-system-cni-dir\") pod \"multus-additional-cni-plugins-s5d7s\" (UID: \"0af3d308-36b9-4622-8544-635f7b24d687\") " pod="openshift-multus/multus-additional-cni-plugins-s5d7s" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.372860 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/0af3d308-36b9-4622-8544-635f7b24d687-cnibin\") pod \"multus-additional-cni-plugins-s5d7s\" (UID: \"0af3d308-36b9-4622-8544-635f7b24d687\") " pod="openshift-multus/multus-additional-cni-plugins-s5d7s" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.372880 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b333a2e0-0da7-4d88-9539-0b4cc12bc2e8-mcd-auth-proxy-config\") pod \"machine-config-daemon-26t88\" (UID: \"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8\") " pod="openshift-machine-config-operator/machine-config-daemon-26t88" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.373027 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/0af3d308-36b9-4622-8544-635f7b24d687-cni-binary-copy\") pod \"multus-additional-cni-plugins-s5d7s\" (UID: \"0af3d308-36b9-4622-8544-635f7b24d687\") " pod="openshift-multus/multus-additional-cni-plugins-s5d7s" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.373083 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/bd007c61-4689-4179-95ad-44a9724e8e7e-ovnkube-script-lib\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.373093 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/0af3d308-36b9-4622-8544-635f7b24d687-os-release\") pod \"multus-additional-cni-plugins-s5d7s\" (UID: \"0af3d308-36b9-4622-8544-635f7b24d687\") " pod="openshift-multus/multus-additional-cni-plugins-s5d7s" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.373123 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/b333a2e0-0da7-4d88-9539-0b4cc12bc2e8-rootfs\") pod \"machine-config-daemon-26t88\" (UID: \"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8\") " pod="openshift-machine-config-operator/machine-config-daemon-26t88" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.373133 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-var-lib-openvswitch\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.373149 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-run-ovn\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.373158 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-host-cni-netd\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.373174 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-host-run-netns\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.373184 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-run-openvswitch\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.373195 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-log-socket\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.373207 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-host-run-ovn-kubernetes\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.373214 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bd007c61-4689-4179-95ad-44a9724e8e7e-ovnkube-config\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.373247 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/0af3d308-36b9-4622-8544-635f7b24d687-tuning-conf-dir\") pod \"multus-additional-cni-plugins-s5d7s\" (UID: \"0af3d308-36b9-4622-8544-635f7b24d687\") " pod="openshift-multus/multus-additional-cni-plugins-s5d7s" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.373273 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-node-log\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.373302 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.373414 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-systemd-units\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.373440 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-host-kubelet\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.373460 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-host-slash\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.373481 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-etc-openvswitch\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.373503 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-host-cni-bin\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.373534 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b333a2e0-0da7-4d88-9539-0b4cc12bc2e8-mcd-auth-proxy-config\") pod \"machine-config-daemon-26t88\" (UID: \"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8\") " pod="openshift-machine-config-operator/machine-config-daemon-26t88" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.373576 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0af3d308-36b9-4622-8544-635f7b24d687-system-cni-dir\") pod \"multus-additional-cni-plugins-s5d7s\" (UID: \"0af3d308-36b9-4622-8544-635f7b24d687\") " pod="openshift-multus/multus-additional-cni-plugins-s5d7s" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.373600 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/0af3d308-36b9-4622-8544-635f7b24d687-cnibin\") pod \"multus-additional-cni-plugins-s5d7s\" (UID: \"0af3d308-36b9-4622-8544-635f7b24d687\") " pod="openshift-multus/multus-additional-cni-plugins-s5d7s" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.373671 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/0af3d308-36b9-4622-8544-635f7b24d687-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-s5d7s\" (UID: \"0af3d308-36b9-4622-8544-635f7b24d687\") " pod="openshift-multus/multus-additional-cni-plugins-s5d7s" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.373930 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bd007c61-4689-4179-95ad-44a9724e8e7e-env-overrides\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.374109 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22ed28a-3748-4eb9-b3a7-c02aa38f7e94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33c0540ccb097b15f65369ff537aef4ae621cc828ba314eb61f0a0037608cadf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c241d29297a73dce12bd0bc2863c56475eb35b8fe8c93991ba0ef8695e38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f42a55d797c6c94af3a5eac1bf08aa38a82359b7ddd0ba32dcc797218ceee6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebed9ec415bb433935da36df8014f9087236cea9788b2048cbff8714f8ad0df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:10Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.376606 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b333a2e0-0da7-4d88-9539-0b4cc12bc2e8-proxy-tls\") pod \"machine-config-daemon-26t88\" (UID: \"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8\") " pod="openshift-machine-config-operator/machine-config-daemon-26t88" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.378232 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bd007c61-4689-4179-95ad-44a9724e8e7e-ovn-node-metrics-cert\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.387244 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpd4j\" (UniqueName: \"kubernetes.io/projected/bd007c61-4689-4179-95ad-44a9724e8e7e-kube-api-access-jpd4j\") pod \"ovnkube-node-fht8l\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.391112 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://deb9f370792598608cf2e0c0fe9975eaf8b95dc94b12eef8c55112df4aa9c998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:10Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.391423 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6v22j\" (UniqueName: \"kubernetes.io/projected/b333a2e0-0da7-4d88-9539-0b4cc12bc2e8-kube-api-access-6v22j\") pod \"machine-config-daemon-26t88\" (UID: \"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8\") " pod="openshift-machine-config-operator/machine-config-daemon-26t88" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.395014 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xlq24\" (UniqueName: \"kubernetes.io/projected/0af3d308-36b9-4622-8544-635f7b24d687-kube-api-access-xlq24\") pod \"multus-additional-cni-plugins-s5d7s\" (UID: \"0af3d308-36b9-4622-8544-635f7b24d687\") " pod="openshift-multus/multus-additional-cni-plugins-s5d7s" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.405297 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:10Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.417736 4799 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-11-29 04:34:09 +0000 UTC, rotation deadline is 2026-09-29 17:41:56.225751966 +0000 UTC Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.417808 4799 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 7309h2m45.807946116s for next certificate rotation Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.423179 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee17b4a9a8999e5f76ceaf69dbd13acf80ed0e95e72f3e1428ce38c5d5982788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://382256e0fa1a4c26610227b9ae96288ccbfbbd97aa3cbd5fd4f18cb4228a234b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:10Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.434333 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt2dl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6c9236-5f7d-46c5-b080-4667620d874e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt2dl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:10Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.445829 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s5d7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0af3d308-36b9-4622-8544-635f7b24d687\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s5d7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:10Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.473368 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.473399 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.473407 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.473422 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.473432 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:10Z","lastTransitionTime":"2025-11-29T04:39:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.528100 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-s5d7s" Nov 29 04:39:10 crc kubenswrapper[4799]: W1129 04:39:10.537585 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0af3d308_36b9_4622_8544_635f7b24d687.slice/crio-b0c8353932b04b7559aabc4b42e186f7bdd981ddcb94a5f4c74b4f42620111ac WatchSource:0}: Error finding container b0c8353932b04b7559aabc4b42e186f7bdd981ddcb94a5f4c74b4f42620111ac: Status 404 returned error can't find the container with id b0c8353932b04b7559aabc4b42e186f7bdd981ddcb94a5f4c74b4f42620111ac Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.541929 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.552163 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-26t88" Nov 29 04:39:10 crc kubenswrapper[4799]: W1129 04:39:10.563841 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb333a2e0_0da7_4d88_9539_0b4cc12bc2e8.slice/crio-cf823e17c0ae1d8457915561e4a721c6ae58dca9a92a79a62d361a3b7d6b7a5d WatchSource:0}: Error finding container cf823e17c0ae1d8457915561e4a721c6ae58dca9a92a79a62d361a3b7d6b7a5d: Status 404 returned error can't find the container with id cf823e17c0ae1d8457915561e4a721c6ae58dca9a92a79a62d361a3b7d6b7a5d Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.576038 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.576065 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.576074 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.576086 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.576095 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:10Z","lastTransitionTime":"2025-11-29T04:39:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.679157 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.679227 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.679238 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.679254 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.679266 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:10Z","lastTransitionTime":"2025-11-29T04:39:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.781734 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.781774 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.781783 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.781815 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.781825 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:10Z","lastTransitionTime":"2025-11-29T04:39:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.809964 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-v7lb4" event={"ID":"55156232-4c80-45a7-aad5-a5fa2793e4ed","Type":"ContainerStarted","Data":"ba6e5473c80737a5b7b00e5874062e0485ed7a4d74632202090a9b12f331953b"} Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.810034 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-v7lb4" event={"ID":"55156232-4c80-45a7-aad5-a5fa2793e4ed","Type":"ContainerStarted","Data":"35868a3fab1626c013aa5cb3cb0b101e91f0acbf3ace3ee04e3998f918dba4ba"} Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.811642 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" event={"ID":"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8","Type":"ContainerStarted","Data":"fd0847696a7b7926047dc62d75739b3b6a7a04cd58efd1c1c2c992d5ae8d20ae"} Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.811677 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" event={"ID":"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8","Type":"ContainerStarted","Data":"83b2ec8f0d384066ba90eb42c1a0a925ade3722b2515b0592d6f2a8f5e533a0f"} Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.811701 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" event={"ID":"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8","Type":"ContainerStarted","Data":"cf823e17c0ae1d8457915561e4a721c6ae58dca9a92a79a62d361a3b7d6b7a5d"} Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.812727 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-qt2dl" event={"ID":"3b6c9236-5f7d-46c5-b080-4667620d874e","Type":"ContainerStarted","Data":"92b6c827d0c57a87808fef301661fce28c50359e85c9e47947c50945f4f869de"} Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.812759 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-qt2dl" event={"ID":"3b6c9236-5f7d-46c5-b080-4667620d874e","Type":"ContainerStarted","Data":"89fca7478f232bd04e23f301ea013cc08349774b081cc2ccd4fccdf6a6e6e99b"} Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.814393 4799 generic.go:334] "Generic (PLEG): container finished" podID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerID="c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64" exitCode=0 Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.814456 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" event={"ID":"bd007c61-4689-4179-95ad-44a9724e8e7e","Type":"ContainerDied","Data":"c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64"} Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.814506 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" event={"ID":"bd007c61-4689-4179-95ad-44a9724e8e7e","Type":"ContainerStarted","Data":"c250bd8866cb517877d111014844a268897f17a961f506bc7259d08e24036a0a"} Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.815700 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-s5d7s" event={"ID":"0af3d308-36b9-4622-8544-635f7b24d687","Type":"ContainerStarted","Data":"75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567"} Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.815732 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-s5d7s" event={"ID":"0af3d308-36b9-4622-8544-635f7b24d687","Type":"ContainerStarted","Data":"b0c8353932b04b7559aabc4b42e186f7bdd981ddcb94a5f4c74b4f42620111ac"} Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.824856 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt2dl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6c9236-5f7d-46c5-b080-4667620d874e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt2dl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:10Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.839097 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s5d7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0af3d308-36b9-4622-8544-635f7b24d687\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s5d7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:10Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.851866 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://deb9f370792598608cf2e0c0fe9975eaf8b95dc94b12eef8c55112df4aa9c998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:10Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.864001 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:10Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.876065 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee17b4a9a8999e5f76ceaf69dbd13acf80ed0e95e72f3e1428ce38c5d5982788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://382256e0fa1a4c26610227b9ae96288ccbfbbd97aa3cbd5fd4f18cb4228a234b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:10Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.883781 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.883841 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.883854 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.883873 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.883885 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:10Z","lastTransitionTime":"2025-11-29T04:39:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.891233 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d40160e-075f-44b7-9d9d-3291e7774f27\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7079548fe9529583a1c129f57838c24cf06d1cb55a97c05c63a05d0cc2e7ec9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7762237003a5a256ff20a16a4869462e2e0dd1426aa391f7ec74ed9b1f572d4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1129 04:39:02.908489 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1129 04:39:02.908637 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 04:39:02.909677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703886220/tls.crt::/tmp/serving-cert-1703886220/tls.key\\\\\\\"\\\\nI1129 04:39:03.077876 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 04:39:03.080368 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 04:39:03.080390 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 04:39:03.080424 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 04:39:03.080432 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 04:39:03.086120 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 04:39:03.086240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086412 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 04:39:03.086502 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 04:39:03.086563 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 04:39:03.086622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1129 04:39:03.086141 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1129 04:39:03.086836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:38:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:10Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.900555 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://396c1c8cbd638c7dc4d8eebded6a17449d9abdc049161578581f1d8c3e31d9a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:10Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.911587 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:10Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.932555 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd007c61-4689-4179-95ad-44a9724e8e7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fht8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:10Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.943189 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-26t88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:10Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.956336 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22ed28a-3748-4eb9-b3a7-c02aa38f7e94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33c0540ccb097b15f65369ff537aef4ae621cc828ba314eb61f0a0037608cadf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c241d29297a73dce12bd0bc2863c56475eb35b8fe8c93991ba0ef8695e38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f42a55d797c6c94af3a5eac1bf08aa38a82359b7ddd0ba32dcc797218ceee6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebed9ec415bb433935da36df8014f9087236cea9788b2048cbff8714f8ad0df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:10Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.968631 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:10Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.977365 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7lb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55156232-4c80-45a7-aad5-a5fa2793e4ed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6e5473c80737a5b7b00e5874062e0485ed7a4d74632202090a9b12f331953b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d2dtw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7lb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:10Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.986467 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.986503 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.986521 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.986537 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.986546 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:10Z","lastTransitionTime":"2025-11-29T04:39:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:10 crc kubenswrapper[4799]: I1129 04:39:10.989195 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22ed28a-3748-4eb9-b3a7-c02aa38f7e94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33c0540ccb097b15f65369ff537aef4ae621cc828ba314eb61f0a0037608cadf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c241d29297a73dce12bd0bc2863c56475eb35b8fe8c93991ba0ef8695e38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f42a55d797c6c94af3a5eac1bf08aa38a82359b7ddd0ba32dcc797218ceee6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebed9ec415bb433935da36df8014f9087236cea9788b2048cbff8714f8ad0df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:10Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.001336 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:10Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.009481 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7lb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55156232-4c80-45a7-aad5-a5fa2793e4ed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6e5473c80737a5b7b00e5874062e0485ed7a4d74632202090a9b12f331953b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d2dtw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7lb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:11Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.019203 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0847696a7b7926047dc62d75739b3b6a7a04cd58efd1c1c2c992d5ae8d20ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b2ec8f0d384066ba90eb42c1a0a925ade3722b2515b0592d6f2a8f5e533a0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-26t88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:11Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.034660 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://deb9f370792598608cf2e0c0fe9975eaf8b95dc94b12eef8c55112df4aa9c998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:11Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.046857 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:11Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.057759 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee17b4a9a8999e5f76ceaf69dbd13acf80ed0e95e72f3e1428ce38c5d5982788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://382256e0fa1a4c26610227b9ae96288ccbfbbd97aa3cbd5fd4f18cb4228a234b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:11Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.068449 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt2dl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6c9236-5f7d-46c5-b080-4667620d874e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b6c827d0c57a87808fef301661fce28c50359e85c9e47947c50945f4f869de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt2dl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:11Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.079770 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s5d7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0af3d308-36b9-4622-8544-635f7b24d687\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s5d7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:11Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.088383 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.088430 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.088443 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.088461 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.088474 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:11Z","lastTransitionTime":"2025-11-29T04:39:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.092564 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d40160e-075f-44b7-9d9d-3291e7774f27\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7079548fe9529583a1c129f57838c24cf06d1cb55a97c05c63a05d0cc2e7ec9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7762237003a5a256ff20a16a4869462e2e0dd1426aa391f7ec74ed9b1f572d4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1129 04:39:02.908489 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1129 04:39:02.908637 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 04:39:02.909677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703886220/tls.crt::/tmp/serving-cert-1703886220/tls.key\\\\\\\"\\\\nI1129 04:39:03.077876 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 04:39:03.080368 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 04:39:03.080390 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 04:39:03.080424 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 04:39:03.080432 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 04:39:03.086120 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 04:39:03.086240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086412 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 04:39:03.086502 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 04:39:03.086563 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 04:39:03.086622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1129 04:39:03.086141 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1129 04:39:03.086836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:38:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:11Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.101981 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://396c1c8cbd638c7dc4d8eebded6a17449d9abdc049161578581f1d8c3e31d9a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:11Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.111544 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:11Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.126596 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd007c61-4689-4179-95ad-44a9724e8e7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fht8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:11Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.190523 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.190564 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.190573 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.190589 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.190600 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:11Z","lastTransitionTime":"2025-11-29T04:39:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.293214 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.293251 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.293260 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.293274 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.293284 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:11Z","lastTransitionTime":"2025-11-29T04:39:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.395760 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.396074 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.396093 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.396117 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.396133 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:11Z","lastTransitionTime":"2025-11-29T04:39:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.484361 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.484458 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.484494 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.484521 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.484550 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:39:11 crc kubenswrapper[4799]: E1129 04:39:11.484621 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 29 04:39:11 crc kubenswrapper[4799]: E1129 04:39:11.484633 4799 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 29 04:39:11 crc kubenswrapper[4799]: E1129 04:39:11.484697 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 29 04:39:11 crc kubenswrapper[4799]: E1129 04:39:11.484719 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-29 04:39:19.484694911 +0000 UTC m=+35.127625321 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 29 04:39:11 crc kubenswrapper[4799]: E1129 04:39:11.484725 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 29 04:39:11 crc kubenswrapper[4799]: E1129 04:39:11.484744 4799 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 04:39:11 crc kubenswrapper[4799]: E1129 04:39:11.484642 4799 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 29 04:39:11 crc kubenswrapper[4799]: E1129 04:39:11.484811 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-29 04:39:19.484778103 +0000 UTC m=+35.127708583 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 04:39:11 crc kubenswrapper[4799]: E1129 04:39:11.484837 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-29 04:39:19.484827285 +0000 UTC m=+35.127757685 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 29 04:39:11 crc kubenswrapper[4799]: E1129 04:39:11.484645 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 29 04:39:11 crc kubenswrapper[4799]: E1129 04:39:11.484853 4799 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 04:39:11 crc kubenswrapper[4799]: E1129 04:39:11.484871 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-29 04:39:19.484866055 +0000 UTC m=+35.127796455 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 04:39:11 crc kubenswrapper[4799]: E1129 04:39:11.485164 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 04:39:19.485146893 +0000 UTC m=+35.128077353 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.497929 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.497963 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.497971 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.497984 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.497993 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:11Z","lastTransitionTime":"2025-11-29T04:39:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.606893 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.607833 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.607848 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.607867 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.607878 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:11Z","lastTransitionTime":"2025-11-29T04:39:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.658660 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 04:39:11 crc kubenswrapper[4799]: E1129 04:39:11.658805 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.658832 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.658840 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:39:11 crc kubenswrapper[4799]: E1129 04:39:11.658959 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 04:39:11 crc kubenswrapper[4799]: E1129 04:39:11.659095 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.710462 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.710499 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.710509 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.710525 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.710535 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:11Z","lastTransitionTime":"2025-11-29T04:39:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.812307 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.812334 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.812343 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.812356 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.812365 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:11Z","lastTransitionTime":"2025-11-29T04:39:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.822106 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" event={"ID":"bd007c61-4689-4179-95ad-44a9724e8e7e","Type":"ContainerStarted","Data":"72802532a47b381ead2507e1ae31c49852f9b1d332f3c08fb1e3a972a128eead"} Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.822150 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" event={"ID":"bd007c61-4689-4179-95ad-44a9724e8e7e","Type":"ContainerStarted","Data":"0c81d94f098efec48c1ff8e172e476bf806ccec5777974eea9c0f3850e5fd40e"} Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.822164 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" event={"ID":"bd007c61-4689-4179-95ad-44a9724e8e7e","Type":"ContainerStarted","Data":"dac1c084955d874741d36709fa848c2c475a2d7108d4378b677bb13a88184105"} Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.822176 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" event={"ID":"bd007c61-4689-4179-95ad-44a9724e8e7e","Type":"ContainerStarted","Data":"28499112c33332c23be39a11cdf7fc4e1946d04524a3f7c534087179a6620121"} Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.822187 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" event={"ID":"bd007c61-4689-4179-95ad-44a9724e8e7e","Type":"ContainerStarted","Data":"e3d7f1ba1a408c54f0470fca40e307b70ba208dc7eeece599b684820e3f56c41"} Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.822221 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" event={"ID":"bd007c61-4689-4179-95ad-44a9724e8e7e","Type":"ContainerStarted","Data":"8e76d91c4f461a1912117cb8ed3a8087c5efaaf0a67455b721bfbeaad1c7dda8"} Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.823911 4799 generic.go:334] "Generic (PLEG): container finished" podID="0af3d308-36b9-4622-8544-635f7b24d687" containerID="75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567" exitCode=0 Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.823965 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-s5d7s" event={"ID":"0af3d308-36b9-4622-8544-635f7b24d687","Type":"ContainerDied","Data":"75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567"} Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.845758 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s5d7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0af3d308-36b9-4622-8544-635f7b24d687\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s5d7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:11Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.865188 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://deb9f370792598608cf2e0c0fe9975eaf8b95dc94b12eef8c55112df4aa9c998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:11Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.877420 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:11Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.889093 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee17b4a9a8999e5f76ceaf69dbd13acf80ed0e95e72f3e1428ce38c5d5982788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://382256e0fa1a4c26610227b9ae96288ccbfbbd97aa3cbd5fd4f18cb4228a234b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:11Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.900692 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt2dl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6c9236-5f7d-46c5-b080-4667620d874e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b6c827d0c57a87808fef301661fce28c50359e85c9e47947c50945f4f869de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt2dl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:11Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.914736 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.914782 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.914818 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.914842 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.914854 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:11Z","lastTransitionTime":"2025-11-29T04:39:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.921521 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d40160e-075f-44b7-9d9d-3291e7774f27\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7079548fe9529583a1c129f57838c24cf06d1cb55a97c05c63a05d0cc2e7ec9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7762237003a5a256ff20a16a4869462e2e0dd1426aa391f7ec74ed9b1f572d4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1129 04:39:02.908489 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1129 04:39:02.908637 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 04:39:02.909677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703886220/tls.crt::/tmp/serving-cert-1703886220/tls.key\\\\\\\"\\\\nI1129 04:39:03.077876 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 04:39:03.080368 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 04:39:03.080390 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 04:39:03.080424 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 04:39:03.080432 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 04:39:03.086120 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 04:39:03.086240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086412 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 04:39:03.086502 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 04:39:03.086563 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 04:39:03.086622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1129 04:39:03.086141 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1129 04:39:03.086836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:38:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:11Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.933176 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://396c1c8cbd638c7dc4d8eebded6a17449d9abdc049161578581f1d8c3e31d9a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:11Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.944149 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:11Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.959732 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd007c61-4689-4179-95ad-44a9724e8e7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fht8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:11Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.970625 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22ed28a-3748-4eb9-b3a7-c02aa38f7e94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33c0540ccb097b15f65369ff537aef4ae621cc828ba314eb61f0a0037608cadf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c241d29297a73dce12bd0bc2863c56475eb35b8fe8c93991ba0ef8695e38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f42a55d797c6c94af3a5eac1bf08aa38a82359b7ddd0ba32dcc797218ceee6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebed9ec415bb433935da36df8014f9087236cea9788b2048cbff8714f8ad0df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:11Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.983268 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:11Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:11 crc kubenswrapper[4799]: I1129 04:39:11.994381 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7lb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55156232-4c80-45a7-aad5-a5fa2793e4ed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6e5473c80737a5b7b00e5874062e0485ed7a4d74632202090a9b12f331953b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d2dtw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7lb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:11Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.006701 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0847696a7b7926047dc62d75739b3b6a7a04cd58efd1c1c2c992d5ae8d20ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b2ec8f0d384066ba90eb42c1a0a925ade3722b2515b0592d6f2a8f5e533a0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-26t88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:12Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.027704 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.027770 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.027814 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.027845 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.027863 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:12Z","lastTransitionTime":"2025-11-29T04:39:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.130222 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.130259 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.130271 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.130288 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.130296 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:12Z","lastTransitionTime":"2025-11-29T04:39:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.231971 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.232007 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.232016 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.232029 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.232170 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:12Z","lastTransitionTime":"2025-11-29T04:39:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.335410 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.335451 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.335462 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.335483 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.335498 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:12Z","lastTransitionTime":"2025-11-29T04:39:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.438070 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.438106 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.438118 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.438132 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.438140 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:12Z","lastTransitionTime":"2025-11-29T04:39:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.451100 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-d87rr"] Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.451411 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-d87rr" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.455055 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.455212 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.455328 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.455244 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.467000 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:12Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.490621 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd007c61-4689-4179-95ad-44a9724e8e7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fht8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:12Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.492824 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6a92ae69-5675-49d9-a567-e2476fe02ef2-host\") pod \"node-ca-d87rr\" (UID: \"6a92ae69-5675-49d9-a567-e2476fe02ef2\") " pod="openshift-image-registry/node-ca-d87rr" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.492873 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffwm2\" (UniqueName: \"kubernetes.io/projected/6a92ae69-5675-49d9-a567-e2476fe02ef2-kube-api-access-ffwm2\") pod \"node-ca-d87rr\" (UID: \"6a92ae69-5675-49d9-a567-e2476fe02ef2\") " pod="openshift-image-registry/node-ca-d87rr" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.492907 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/6a92ae69-5675-49d9-a567-e2476fe02ef2-serviceca\") pod \"node-ca-d87rr\" (UID: \"6a92ae69-5675-49d9-a567-e2476fe02ef2\") " pod="openshift-image-registry/node-ca-d87rr" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.503576 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7lb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55156232-4c80-45a7-aad5-a5fa2793e4ed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6e5473c80737a5b7b00e5874062e0485ed7a4d74632202090a9b12f331953b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d2dtw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7lb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:12Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.516125 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0847696a7b7926047dc62d75739b3b6a7a04cd58efd1c1c2c992d5ae8d20ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b2ec8f0d384066ba90eb42c1a0a925ade3722b2515b0592d6f2a8f5e533a0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-26t88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:12Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.529621 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22ed28a-3748-4eb9-b3a7-c02aa38f7e94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33c0540ccb097b15f65369ff537aef4ae621cc828ba314eb61f0a0037608cadf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c241d29297a73dce12bd0bc2863c56475eb35b8fe8c93991ba0ef8695e38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f42a55d797c6c94af3a5eac1bf08aa38a82359b7ddd0ba32dcc797218ceee6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebed9ec415bb433935da36df8014f9087236cea9788b2048cbff8714f8ad0df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:12Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.540437 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.540488 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.540497 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.540514 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.540526 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:12Z","lastTransitionTime":"2025-11-29T04:39:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.542879 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:12Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.555531 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee17b4a9a8999e5f76ceaf69dbd13acf80ed0e95e72f3e1428ce38c5d5982788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://382256e0fa1a4c26610227b9ae96288ccbfbbd97aa3cbd5fd4f18cb4228a234b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:12Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.568668 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt2dl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6c9236-5f7d-46c5-b080-4667620d874e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b6c827d0c57a87808fef301661fce28c50359e85c9e47947c50945f4f869de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt2dl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:12Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.584363 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s5d7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0af3d308-36b9-4622-8544-635f7b24d687\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s5d7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:12Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.592984 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d87rr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a92ae69-5675-49d9-a567-e2476fe02ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffwm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d87rr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:12Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.593489 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6a92ae69-5675-49d9-a567-e2476fe02ef2-host\") pod \"node-ca-d87rr\" (UID: \"6a92ae69-5675-49d9-a567-e2476fe02ef2\") " pod="openshift-image-registry/node-ca-d87rr" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.593631 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffwm2\" (UniqueName: \"kubernetes.io/projected/6a92ae69-5675-49d9-a567-e2476fe02ef2-kube-api-access-ffwm2\") pod \"node-ca-d87rr\" (UID: \"6a92ae69-5675-49d9-a567-e2476fe02ef2\") " pod="openshift-image-registry/node-ca-d87rr" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.593583 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6a92ae69-5675-49d9-a567-e2476fe02ef2-host\") pod \"node-ca-d87rr\" (UID: \"6a92ae69-5675-49d9-a567-e2476fe02ef2\") " pod="openshift-image-registry/node-ca-d87rr" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.593736 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/6a92ae69-5675-49d9-a567-e2476fe02ef2-serviceca\") pod \"node-ca-d87rr\" (UID: \"6a92ae69-5675-49d9-a567-e2476fe02ef2\") " pod="openshift-image-registry/node-ca-d87rr" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.594892 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/6a92ae69-5675-49d9-a567-e2476fe02ef2-serviceca\") pod \"node-ca-d87rr\" (UID: \"6a92ae69-5675-49d9-a567-e2476fe02ef2\") " pod="openshift-image-registry/node-ca-d87rr" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.604414 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://deb9f370792598608cf2e0c0fe9975eaf8b95dc94b12eef8c55112df4aa9c998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:12Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.615085 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffwm2\" (UniqueName: \"kubernetes.io/projected/6a92ae69-5675-49d9-a567-e2476fe02ef2-kube-api-access-ffwm2\") pod \"node-ca-d87rr\" (UID: \"6a92ae69-5675-49d9-a567-e2476fe02ef2\") " pod="openshift-image-registry/node-ca-d87rr" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.618833 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:12Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.632947 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d40160e-075f-44b7-9d9d-3291e7774f27\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7079548fe9529583a1c129f57838c24cf06d1cb55a97c05c63a05d0cc2e7ec9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7762237003a5a256ff20a16a4869462e2e0dd1426aa391f7ec74ed9b1f572d4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1129 04:39:02.908489 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1129 04:39:02.908637 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 04:39:02.909677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703886220/tls.crt::/tmp/serving-cert-1703886220/tls.key\\\\\\\"\\\\nI1129 04:39:03.077876 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 04:39:03.080368 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 04:39:03.080390 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 04:39:03.080424 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 04:39:03.080432 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 04:39:03.086120 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 04:39:03.086240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086412 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 04:39:03.086502 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 04:39:03.086563 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 04:39:03.086622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1129 04:39:03.086141 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1129 04:39:03.086836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:38:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:12Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.643276 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.643320 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.643331 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.643351 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.643364 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:12Z","lastTransitionTime":"2025-11-29T04:39:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.648212 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://396c1c8cbd638c7dc4d8eebded6a17449d9abdc049161578581f1d8c3e31d9a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:12Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.745900 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.745933 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.745941 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.745956 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.745968 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:12Z","lastTransitionTime":"2025-11-29T04:39:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.764137 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-d87rr" Nov 29 04:39:12 crc kubenswrapper[4799]: W1129 04:39:12.775653 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6a92ae69_5675_49d9_a567_e2476fe02ef2.slice/crio-6c27441b9265ed5e38c982b06a981f8f9d9fa3673e942f1aaf67d80a2c3f156e WatchSource:0}: Error finding container 6c27441b9265ed5e38c982b06a981f8f9d9fa3673e942f1aaf67d80a2c3f156e: Status 404 returned error can't find the container with id 6c27441b9265ed5e38c982b06a981f8f9d9fa3673e942f1aaf67d80a2c3f156e Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.828053 4799 generic.go:334] "Generic (PLEG): container finished" podID="0af3d308-36b9-4622-8544-635f7b24d687" containerID="cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d" exitCode=0 Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.828100 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-s5d7s" event={"ID":"0af3d308-36b9-4622-8544-635f7b24d687","Type":"ContainerDied","Data":"cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d"} Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.831157 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-d87rr" event={"ID":"6a92ae69-5675-49d9-a567-e2476fe02ef2","Type":"ContainerStarted","Data":"6c27441b9265ed5e38c982b06a981f8f9d9fa3673e942f1aaf67d80a2c3f156e"} Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.841356 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://deb9f370792598608cf2e0c0fe9975eaf8b95dc94b12eef8c55112df4aa9c998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:12Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.848374 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.848412 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.848425 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.848443 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.848455 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:12Z","lastTransitionTime":"2025-11-29T04:39:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.854926 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:12Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.867622 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee17b4a9a8999e5f76ceaf69dbd13acf80ed0e95e72f3e1428ce38c5d5982788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://382256e0fa1a4c26610227b9ae96288ccbfbbd97aa3cbd5fd4f18cb4228a234b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:12Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.880167 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt2dl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6c9236-5f7d-46c5-b080-4667620d874e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b6c827d0c57a87808fef301661fce28c50359e85c9e47947c50945f4f869de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt2dl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:12Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.895151 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s5d7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0af3d308-36b9-4622-8544-635f7b24d687\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s5d7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:12Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.906631 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d87rr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a92ae69-5675-49d9-a567-e2476fe02ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffwm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d87rr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:12Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.922682 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d40160e-075f-44b7-9d9d-3291e7774f27\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7079548fe9529583a1c129f57838c24cf06d1cb55a97c05c63a05d0cc2e7ec9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7762237003a5a256ff20a16a4869462e2e0dd1426aa391f7ec74ed9b1f572d4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1129 04:39:02.908489 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1129 04:39:02.908637 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 04:39:02.909677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703886220/tls.crt::/tmp/serving-cert-1703886220/tls.key\\\\\\\"\\\\nI1129 04:39:03.077876 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 04:39:03.080368 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 04:39:03.080390 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 04:39:03.080424 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 04:39:03.080432 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 04:39:03.086120 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 04:39:03.086240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086412 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 04:39:03.086502 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 04:39:03.086563 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 04:39:03.086622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1129 04:39:03.086141 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1129 04:39:03.086836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:38:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:12Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.936307 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://396c1c8cbd638c7dc4d8eebded6a17449d9abdc049161578581f1d8c3e31d9a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:12Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.950656 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:12Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.951911 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.952004 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.952015 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.952028 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.952069 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:12Z","lastTransitionTime":"2025-11-29T04:39:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.971857 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd007c61-4689-4179-95ad-44a9724e8e7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fht8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:12Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.985299 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22ed28a-3748-4eb9-b3a7-c02aa38f7e94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33c0540ccb097b15f65369ff537aef4ae621cc828ba314eb61f0a0037608cadf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c241d29297a73dce12bd0bc2863c56475eb35b8fe8c93991ba0ef8695e38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f42a55d797c6c94af3a5eac1bf08aa38a82359b7ddd0ba32dcc797218ceee6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebed9ec415bb433935da36df8014f9087236cea9788b2048cbff8714f8ad0df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:12Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:12 crc kubenswrapper[4799]: I1129 04:39:12.995920 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:12Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.005467 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7lb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55156232-4c80-45a7-aad5-a5fa2793e4ed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6e5473c80737a5b7b00e5874062e0485ed7a4d74632202090a9b12f331953b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d2dtw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7lb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:13Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.016362 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0847696a7b7926047dc62d75739b3b6a7a04cd58efd1c1c2c992d5ae8d20ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b2ec8f0d384066ba90eb42c1a0a925ade3722b2515b0592d6f2a8f5e533a0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-26t88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:13Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.054174 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.054209 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.054218 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.054232 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.054241 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:13Z","lastTransitionTime":"2025-11-29T04:39:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.156701 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.156746 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.156760 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.156778 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.156816 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:13Z","lastTransitionTime":"2025-11-29T04:39:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.259972 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.260024 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.260037 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.260057 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.260071 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:13Z","lastTransitionTime":"2025-11-29T04:39:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.362623 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.362663 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.362672 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.362686 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.362696 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:13Z","lastTransitionTime":"2025-11-29T04:39:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.464806 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.464848 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.464858 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.464877 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.464887 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:13Z","lastTransitionTime":"2025-11-29T04:39:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.567955 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.567990 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.567999 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.568013 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.568025 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:13Z","lastTransitionTime":"2025-11-29T04:39:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.658088 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.658155 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.658097 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 04:39:13 crc kubenswrapper[4799]: E1129 04:39:13.658207 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 04:39:13 crc kubenswrapper[4799]: E1129 04:39:13.658345 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 04:39:13 crc kubenswrapper[4799]: E1129 04:39:13.658479 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.670271 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.670305 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.670317 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.670333 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.670344 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:13Z","lastTransitionTime":"2025-11-29T04:39:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.772605 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.772670 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.772693 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.772722 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.772744 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:13Z","lastTransitionTime":"2025-11-29T04:39:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.834759 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-d87rr" event={"ID":"6a92ae69-5675-49d9-a567-e2476fe02ef2","Type":"ContainerStarted","Data":"207d152294f9996ed7294d858842da5eef891a6d8a027473dc3d29a7101aadab"} Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.837311 4799 generic.go:334] "Generic (PLEG): container finished" podID="0af3d308-36b9-4622-8544-635f7b24d687" containerID="3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46" exitCode=0 Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.837395 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-s5d7s" event={"ID":"0af3d308-36b9-4622-8544-635f7b24d687","Type":"ContainerDied","Data":"3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46"} Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.853184 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d40160e-075f-44b7-9d9d-3291e7774f27\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7079548fe9529583a1c129f57838c24cf06d1cb55a97c05c63a05d0cc2e7ec9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7762237003a5a256ff20a16a4869462e2e0dd1426aa391f7ec74ed9b1f572d4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1129 04:39:02.908489 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1129 04:39:02.908637 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 04:39:02.909677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703886220/tls.crt::/tmp/serving-cert-1703886220/tls.key\\\\\\\"\\\\nI1129 04:39:03.077876 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 04:39:03.080368 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 04:39:03.080390 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 04:39:03.080424 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 04:39:03.080432 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 04:39:03.086120 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 04:39:03.086240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086412 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 04:39:03.086502 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 04:39:03.086563 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 04:39:03.086622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1129 04:39:03.086141 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1129 04:39:03.086836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:38:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:13Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.867620 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://396c1c8cbd638c7dc4d8eebded6a17449d9abdc049161578581f1d8c3e31d9a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:13Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.875134 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.875172 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.875182 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.875209 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.875226 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:13Z","lastTransitionTime":"2025-11-29T04:39:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.883155 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:13Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.912325 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd007c61-4689-4179-95ad-44a9724e8e7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fht8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:13Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.930925 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0847696a7b7926047dc62d75739b3b6a7a04cd58efd1c1c2c992d5ae8d20ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b2ec8f0d384066ba90eb42c1a0a925ade3722b2515b0592d6f2a8f5e533a0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-26t88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:13Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.944237 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22ed28a-3748-4eb9-b3a7-c02aa38f7e94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33c0540ccb097b15f65369ff537aef4ae621cc828ba314eb61f0a0037608cadf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c241d29297a73dce12bd0bc2863c56475eb35b8fe8c93991ba0ef8695e38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f42a55d797c6c94af3a5eac1bf08aa38a82359b7ddd0ba32dcc797218ceee6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebed9ec415bb433935da36df8014f9087236cea9788b2048cbff8714f8ad0df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:13Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.957491 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:13Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.968710 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7lb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55156232-4c80-45a7-aad5-a5fa2793e4ed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6e5473c80737a5b7b00e5874062e0485ed7a4d74632202090a9b12f331953b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d2dtw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7lb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:13Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.977862 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.977912 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.977929 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.977952 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.977970 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:13Z","lastTransitionTime":"2025-11-29T04:39:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.984641 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt2dl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6c9236-5f7d-46c5-b080-4667620d874e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b6c827d0c57a87808fef301661fce28c50359e85c9e47947c50945f4f869de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt2dl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:13Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:13 crc kubenswrapper[4799]: I1129 04:39:13.998636 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s5d7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0af3d308-36b9-4622-8544-635f7b24d687\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s5d7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:13Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.008448 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d87rr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a92ae69-5675-49d9-a567-e2476fe02ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://207d152294f9996ed7294d858842da5eef891a6d8a027473dc3d29a7101aadab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffwm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d87rr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:14Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.021206 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://deb9f370792598608cf2e0c0fe9975eaf8b95dc94b12eef8c55112df4aa9c998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:14Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.033297 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:14Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.043233 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee17b4a9a8999e5f76ceaf69dbd13acf80ed0e95e72f3e1428ce38c5d5982788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://382256e0fa1a4c26610227b9ae96288ccbfbbd97aa3cbd5fd4f18cb4228a234b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:14Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.055645 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:14Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.071545 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd007c61-4689-4179-95ad-44a9724e8e7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fht8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:14Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.079858 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.079891 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.079901 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.079916 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.079926 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:14Z","lastTransitionTime":"2025-11-29T04:39:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.084139 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22ed28a-3748-4eb9-b3a7-c02aa38f7e94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33c0540ccb097b15f65369ff537aef4ae621cc828ba314eb61f0a0037608cadf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c241d29297a73dce12bd0bc2863c56475eb35b8fe8c93991ba0ef8695e38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f42a55d797c6c94af3a5eac1bf08aa38a82359b7ddd0ba32dcc797218ceee6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebed9ec415bb433935da36df8014f9087236cea9788b2048cbff8714f8ad0df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:14Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.094321 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:14Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.103509 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7lb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55156232-4c80-45a7-aad5-a5fa2793e4ed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6e5473c80737a5b7b00e5874062e0485ed7a4d74632202090a9b12f331953b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d2dtw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7lb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:14Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.112573 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0847696a7b7926047dc62d75739b3b6a7a04cd58efd1c1c2c992d5ae8d20ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b2ec8f0d384066ba90eb42c1a0a925ade3722b2515b0592d6f2a8f5e533a0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-26t88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:14Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.123376 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://deb9f370792598608cf2e0c0fe9975eaf8b95dc94b12eef8c55112df4aa9c998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:14Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.134526 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:14Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.144317 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee17b4a9a8999e5f76ceaf69dbd13acf80ed0e95e72f3e1428ce38c5d5982788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://382256e0fa1a4c26610227b9ae96288ccbfbbd97aa3cbd5fd4f18cb4228a234b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:14Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.156632 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt2dl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6c9236-5f7d-46c5-b080-4667620d874e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b6c827d0c57a87808fef301661fce28c50359e85c9e47947c50945f4f869de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt2dl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:14Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.168953 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s5d7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0af3d308-36b9-4622-8544-635f7b24d687\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s5d7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:14Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.176991 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d87rr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a92ae69-5675-49d9-a567-e2476fe02ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://207d152294f9996ed7294d858842da5eef891a6d8a027473dc3d29a7101aadab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffwm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d87rr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:14Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.182505 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.182548 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.182564 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.182580 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.182591 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:14Z","lastTransitionTime":"2025-11-29T04:39:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.188445 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d40160e-075f-44b7-9d9d-3291e7774f27\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7079548fe9529583a1c129f57838c24cf06d1cb55a97c05c63a05d0cc2e7ec9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7762237003a5a256ff20a16a4869462e2e0dd1426aa391f7ec74ed9b1f572d4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1129 04:39:02.908489 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1129 04:39:02.908637 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 04:39:02.909677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703886220/tls.crt::/tmp/serving-cert-1703886220/tls.key\\\\\\\"\\\\nI1129 04:39:03.077876 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 04:39:03.080368 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 04:39:03.080390 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 04:39:03.080424 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 04:39:03.080432 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 04:39:03.086120 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 04:39:03.086240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086412 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 04:39:03.086502 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 04:39:03.086563 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 04:39:03.086622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1129 04:39:03.086141 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1129 04:39:03.086836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:38:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:14Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.198191 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://396c1c8cbd638c7dc4d8eebded6a17449d9abdc049161578581f1d8c3e31d9a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:14Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.285043 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.285076 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.285085 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.285097 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.285106 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:14Z","lastTransitionTime":"2025-11-29T04:39:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.387646 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.387674 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.387694 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.387708 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.387718 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:14Z","lastTransitionTime":"2025-11-29T04:39:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.490856 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.491220 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.491237 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.491262 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.491279 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:14Z","lastTransitionTime":"2025-11-29T04:39:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.515544 4799 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.594004 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.594071 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.594095 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.594124 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.594147 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:14Z","lastTransitionTime":"2025-11-29T04:39:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.673766 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:14Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.692346 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd007c61-4689-4179-95ad-44a9724e8e7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fht8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:14Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.697122 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.697192 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.697219 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.697250 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.697271 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:14Z","lastTransitionTime":"2025-11-29T04:39:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.710747 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22ed28a-3748-4eb9-b3a7-c02aa38f7e94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33c0540ccb097b15f65369ff537aef4ae621cc828ba314eb61f0a0037608cadf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c241d29297a73dce12bd0bc2863c56475eb35b8fe8c93991ba0ef8695e38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f42a55d797c6c94af3a5eac1bf08aa38a82359b7ddd0ba32dcc797218ceee6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebed9ec415bb433935da36df8014f9087236cea9788b2048cbff8714f8ad0df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:14Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.729009 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:14Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.740979 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7lb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55156232-4c80-45a7-aad5-a5fa2793e4ed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6e5473c80737a5b7b00e5874062e0485ed7a4d74632202090a9b12f331953b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d2dtw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7lb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:14Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.756605 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0847696a7b7926047dc62d75739b3b6a7a04cd58efd1c1c2c992d5ae8d20ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b2ec8f0d384066ba90eb42c1a0a925ade3722b2515b0592d6f2a8f5e533a0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-26t88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:14Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.769659 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://deb9f370792598608cf2e0c0fe9975eaf8b95dc94b12eef8c55112df4aa9c998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:14Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.782132 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:14Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.800056 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.800091 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.800100 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.800115 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.800125 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:14Z","lastTransitionTime":"2025-11-29T04:39:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.801301 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee17b4a9a8999e5f76ceaf69dbd13acf80ed0e95e72f3e1428ce38c5d5982788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://382256e0fa1a4c26610227b9ae96288ccbfbbd97aa3cbd5fd4f18cb4228a234b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:14Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.817327 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt2dl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6c9236-5f7d-46c5-b080-4667620d874e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b6c827d0c57a87808fef301661fce28c50359e85c9e47947c50945f4f869de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt2dl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:14Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.839518 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s5d7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0af3d308-36b9-4622-8544-635f7b24d687\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s5d7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:14Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.845458 4799 generic.go:334] "Generic (PLEG): container finished" podID="0af3d308-36b9-4622-8544-635f7b24d687" containerID="0088b60069e830e9c36075b09d1ff135da79e94fe6cd271d0e93415af50557e0" exitCode=0 Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.845512 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-s5d7s" event={"ID":"0af3d308-36b9-4622-8544-635f7b24d687","Type":"ContainerDied","Data":"0088b60069e830e9c36075b09d1ff135da79e94fe6cd271d0e93415af50557e0"} Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.849731 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" event={"ID":"bd007c61-4689-4179-95ad-44a9724e8e7e","Type":"ContainerStarted","Data":"d135e42b8886f1f118abf8f4ce4cf2875ac4dc6c148bfc0677cef6b3651562c0"} Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.855559 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d87rr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a92ae69-5675-49d9-a567-e2476fe02ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://207d152294f9996ed7294d858842da5eef891a6d8a027473dc3d29a7101aadab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffwm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d87rr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:14Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.870332 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d40160e-075f-44b7-9d9d-3291e7774f27\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7079548fe9529583a1c129f57838c24cf06d1cb55a97c05c63a05d0cc2e7ec9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7762237003a5a256ff20a16a4869462e2e0dd1426aa391f7ec74ed9b1f572d4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1129 04:39:02.908489 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1129 04:39:02.908637 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 04:39:02.909677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703886220/tls.crt::/tmp/serving-cert-1703886220/tls.key\\\\\\\"\\\\nI1129 04:39:03.077876 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 04:39:03.080368 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 04:39:03.080390 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 04:39:03.080424 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 04:39:03.080432 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 04:39:03.086120 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 04:39:03.086240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086412 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 04:39:03.086502 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 04:39:03.086563 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 04:39:03.086622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1129 04:39:03.086141 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1129 04:39:03.086836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:38:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:14Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.884478 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://396c1c8cbd638c7dc4d8eebded6a17449d9abdc049161578581f1d8c3e31d9a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:14Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.896669 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:14Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.905342 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.905389 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.905400 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.905419 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.905433 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:14Z","lastTransitionTime":"2025-11-29T04:39:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.960454 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd007c61-4689-4179-95ad-44a9724e8e7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fht8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:14Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.970133 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7lb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55156232-4c80-45a7-aad5-a5fa2793e4ed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6e5473c80737a5b7b00e5874062e0485ed7a4d74632202090a9b12f331953b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d2dtw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7lb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:14Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.980313 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0847696a7b7926047dc62d75739b3b6a7a04cd58efd1c1c2c992d5ae8d20ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b2ec8f0d384066ba90eb42c1a0a925ade3722b2515b0592d6f2a8f5e533a0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-26t88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:14Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:14 crc kubenswrapper[4799]: I1129 04:39:14.992426 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22ed28a-3748-4eb9-b3a7-c02aa38f7e94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33c0540ccb097b15f65369ff537aef4ae621cc828ba314eb61f0a0037608cadf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c241d29297a73dce12bd0bc2863c56475eb35b8fe8c93991ba0ef8695e38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f42a55d797c6c94af3a5eac1bf08aa38a82359b7ddd0ba32dcc797218ceee6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebed9ec415bb433935da36df8014f9087236cea9788b2048cbff8714f8ad0df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:14Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.003521 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:15Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.008254 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.008284 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.008293 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.008305 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.008315 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:15Z","lastTransitionTime":"2025-11-29T04:39:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.019345 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee17b4a9a8999e5f76ceaf69dbd13acf80ed0e95e72f3e1428ce38c5d5982788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://382256e0fa1a4c26610227b9ae96288ccbfbbd97aa3cbd5fd4f18cb4228a234b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:15Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.030866 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt2dl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6c9236-5f7d-46c5-b080-4667620d874e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b6c827d0c57a87808fef301661fce28c50359e85c9e47947c50945f4f869de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt2dl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:15Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.045842 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s5d7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0af3d308-36b9-4622-8544-635f7b24d687\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0088b60069e830e9c36075b09d1ff135da79e94fe6cd271d0e93415af50557e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0088b60069e830e9c36075b09d1ff135da79e94fe6cd271d0e93415af50557e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s5d7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:15Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.055605 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d87rr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a92ae69-5675-49d9-a567-e2476fe02ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://207d152294f9996ed7294d858842da5eef891a6d8a027473dc3d29a7101aadab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffwm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d87rr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:15Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.067566 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://deb9f370792598608cf2e0c0fe9975eaf8b95dc94b12eef8c55112df4aa9c998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:15Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.080311 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:15Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.093672 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d40160e-075f-44b7-9d9d-3291e7774f27\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7079548fe9529583a1c129f57838c24cf06d1cb55a97c05c63a05d0cc2e7ec9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7762237003a5a256ff20a16a4869462e2e0dd1426aa391f7ec74ed9b1f572d4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1129 04:39:02.908489 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1129 04:39:02.908637 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 04:39:02.909677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703886220/tls.crt::/tmp/serving-cert-1703886220/tls.key\\\\\\\"\\\\nI1129 04:39:03.077876 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 04:39:03.080368 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 04:39:03.080390 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 04:39:03.080424 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 04:39:03.080432 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 04:39:03.086120 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 04:39:03.086240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086412 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 04:39:03.086502 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 04:39:03.086563 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 04:39:03.086622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1129 04:39:03.086141 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1129 04:39:03.086836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:38:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:15Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.105579 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://396c1c8cbd638c7dc4d8eebded6a17449d9abdc049161578581f1d8c3e31d9a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:15Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.109985 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.110072 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.110083 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.110100 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.110109 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:15Z","lastTransitionTime":"2025-11-29T04:39:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.211920 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.211952 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.211962 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.211975 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.211983 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:15Z","lastTransitionTime":"2025-11-29T04:39:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.314900 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.314935 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.314947 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.314960 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.314970 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:15Z","lastTransitionTime":"2025-11-29T04:39:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.418365 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.418402 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.418411 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.418427 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.418439 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:15Z","lastTransitionTime":"2025-11-29T04:39:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.520832 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.520863 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.520874 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.520890 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.520900 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:15Z","lastTransitionTime":"2025-11-29T04:39:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.623734 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.623768 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.623776 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.623808 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.623818 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:15Z","lastTransitionTime":"2025-11-29T04:39:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.658170 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.658202 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:39:15 crc kubenswrapper[4799]: E1129 04:39:15.658265 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.658222 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 04:39:15 crc kubenswrapper[4799]: E1129 04:39:15.658373 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 04:39:15 crc kubenswrapper[4799]: E1129 04:39:15.658448 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.726667 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.726748 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.726807 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.726827 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.726837 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:15Z","lastTransitionTime":"2025-11-29T04:39:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.828640 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.828693 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.828706 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.828721 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.828733 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:15Z","lastTransitionTime":"2025-11-29T04:39:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.855962 4799 generic.go:334] "Generic (PLEG): container finished" podID="0af3d308-36b9-4622-8544-635f7b24d687" containerID="fd9b0b72fbe8b552d1266c310f6a6a67822c4c70dec1f812d31cc00748d50565" exitCode=0 Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.856018 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-s5d7s" event={"ID":"0af3d308-36b9-4622-8544-635f7b24d687","Type":"ContainerDied","Data":"fd9b0b72fbe8b552d1266c310f6a6a67822c4c70dec1f812d31cc00748d50565"} Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.872269 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:15Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.906563 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd007c61-4689-4179-95ad-44a9724e8e7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fht8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:15Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.922342 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22ed28a-3748-4eb9-b3a7-c02aa38f7e94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33c0540ccb097b15f65369ff537aef4ae621cc828ba314eb61f0a0037608cadf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c241d29297a73dce12bd0bc2863c56475eb35b8fe8c93991ba0ef8695e38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f42a55d797c6c94af3a5eac1bf08aa38a82359b7ddd0ba32dcc797218ceee6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebed9ec415bb433935da36df8014f9087236cea9788b2048cbff8714f8ad0df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:15Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.931486 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.931536 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.931551 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.931575 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.931591 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:15Z","lastTransitionTime":"2025-11-29T04:39:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.935336 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:15Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.947760 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7lb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55156232-4c80-45a7-aad5-a5fa2793e4ed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6e5473c80737a5b7b00e5874062e0485ed7a4d74632202090a9b12f331953b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d2dtw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7lb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:15Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.961154 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0847696a7b7926047dc62d75739b3b6a7a04cd58efd1c1c2c992d5ae8d20ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b2ec8f0d384066ba90eb42c1a0a925ade3722b2515b0592d6f2a8f5e533a0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-26t88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:15Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.974900 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://deb9f370792598608cf2e0c0fe9975eaf8b95dc94b12eef8c55112df4aa9c998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:15Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:15 crc kubenswrapper[4799]: I1129 04:39:15.989060 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:15Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.003681 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee17b4a9a8999e5f76ceaf69dbd13acf80ed0e95e72f3e1428ce38c5d5982788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://382256e0fa1a4c26610227b9ae96288ccbfbbd97aa3cbd5fd4f18cb4228a234b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:16Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.015164 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt2dl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6c9236-5f7d-46c5-b080-4667620d874e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b6c827d0c57a87808fef301661fce28c50359e85c9e47947c50945f4f869de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt2dl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:16Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.029109 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s5d7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0af3d308-36b9-4622-8544-635f7b24d687\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0088b60069e830e9c36075b09d1ff135da79e94fe6cd271d0e93415af50557e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0088b60069e830e9c36075b09d1ff135da79e94fe6cd271d0e93415af50557e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9b0b72fbe8b552d1266c310f6a6a67822c4c70dec1f812d31cc00748d50565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9b0b72fbe8b552d1266c310f6a6a67822c4c70dec1f812d31cc00748d50565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s5d7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:16Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.033867 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.033901 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.033912 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.033928 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.033942 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:16Z","lastTransitionTime":"2025-11-29T04:39:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.039327 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d87rr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a92ae69-5675-49d9-a567-e2476fe02ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://207d152294f9996ed7294d858842da5eef891a6d8a027473dc3d29a7101aadab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffwm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d87rr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:16Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.052862 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d40160e-075f-44b7-9d9d-3291e7774f27\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7079548fe9529583a1c129f57838c24cf06d1cb55a97c05c63a05d0cc2e7ec9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7762237003a5a256ff20a16a4869462e2e0dd1426aa391f7ec74ed9b1f572d4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1129 04:39:02.908489 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1129 04:39:02.908637 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 04:39:02.909677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703886220/tls.crt::/tmp/serving-cert-1703886220/tls.key\\\\\\\"\\\\nI1129 04:39:03.077876 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 04:39:03.080368 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 04:39:03.080390 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 04:39:03.080424 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 04:39:03.080432 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 04:39:03.086120 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 04:39:03.086240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086412 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 04:39:03.086502 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 04:39:03.086563 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 04:39:03.086622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1129 04:39:03.086141 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1129 04:39:03.086836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:38:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:16Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.064483 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://396c1c8cbd638c7dc4d8eebded6a17449d9abdc049161578581f1d8c3e31d9a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:16Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.135853 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.135894 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.135906 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.135923 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.135935 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:16Z","lastTransitionTime":"2025-11-29T04:39:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.238291 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.238336 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.238348 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.238367 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.238380 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:16Z","lastTransitionTime":"2025-11-29T04:39:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.340442 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.340492 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.340510 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.340527 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.340540 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:16Z","lastTransitionTime":"2025-11-29T04:39:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.443103 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.443173 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.443243 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.443268 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.443284 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:16Z","lastTransitionTime":"2025-11-29T04:39:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.546258 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.546299 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.546311 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.546327 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.546337 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:16Z","lastTransitionTime":"2025-11-29T04:39:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.649767 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.649845 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.649861 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.649883 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.649897 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:16Z","lastTransitionTime":"2025-11-29T04:39:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.752098 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.752143 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.752156 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.752174 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.752186 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:16Z","lastTransitionTime":"2025-11-29T04:39:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.854834 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.854868 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.854876 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.854891 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.854900 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:16Z","lastTransitionTime":"2025-11-29T04:39:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.863006 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" event={"ID":"bd007c61-4689-4179-95ad-44a9724e8e7e","Type":"ContainerStarted","Data":"ea607ff09f683e7ce5d058b6b45b53cf355a787bed536da93bc6d52dac1334e7"} Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.863317 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.863383 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.868552 4799 generic.go:334] "Generic (PLEG): container finished" podID="0af3d308-36b9-4622-8544-635f7b24d687" containerID="aa3b85ae7c47440f0f61436e1cb1f3f2096e0e4011434876f9d17e913ca24df7" exitCode=0 Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.868612 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-s5d7s" event={"ID":"0af3d308-36b9-4622-8544-635f7b24d687","Type":"ContainerDied","Data":"aa3b85ae7c47440f0f61436e1cb1f3f2096e0e4011434876f9d17e913ca24df7"} Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.885081 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22ed28a-3748-4eb9-b3a7-c02aa38f7e94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33c0540ccb097b15f65369ff537aef4ae621cc828ba314eb61f0a0037608cadf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c241d29297a73dce12bd0bc2863c56475eb35b8fe8c93991ba0ef8695e38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f42a55d797c6c94af3a5eac1bf08aa38a82359b7ddd0ba32dcc797218ceee6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebed9ec415bb433935da36df8014f9087236cea9788b2048cbff8714f8ad0df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:16Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.939564 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.940169 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.940865 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:16Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.957333 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7lb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55156232-4c80-45a7-aad5-a5fa2793e4ed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6e5473c80737a5b7b00e5874062e0485ed7a4d74632202090a9b12f331953b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d2dtw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7lb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:16Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.958126 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.958182 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.958200 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.958222 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.958240 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:16Z","lastTransitionTime":"2025-11-29T04:39:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.974844 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0847696a7b7926047dc62d75739b3b6a7a04cd58efd1c1c2c992d5ae8d20ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b2ec8f0d384066ba90eb42c1a0a925ade3722b2515b0592d6f2a8f5e533a0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-26t88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:16Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:16 crc kubenswrapper[4799]: I1129 04:39:16.991131 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://deb9f370792598608cf2e0c0fe9975eaf8b95dc94b12eef8c55112df4aa9c998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:16Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.009362 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:17Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.026176 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee17b4a9a8999e5f76ceaf69dbd13acf80ed0e95e72f3e1428ce38c5d5982788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://382256e0fa1a4c26610227b9ae96288ccbfbbd97aa3cbd5fd4f18cb4228a234b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:17Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.045395 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt2dl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6c9236-5f7d-46c5-b080-4667620d874e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b6c827d0c57a87808fef301661fce28c50359e85c9e47947c50945f4f869de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt2dl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:17Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.061582 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.061624 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.061636 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.061658 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.061671 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:17Z","lastTransitionTime":"2025-11-29T04:39:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.063825 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s5d7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0af3d308-36b9-4622-8544-635f7b24d687\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0088b60069e830e9c36075b09d1ff135da79e94fe6cd271d0e93415af50557e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0088b60069e830e9c36075b09d1ff135da79e94fe6cd271d0e93415af50557e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9b0b72fbe8b552d1266c310f6a6a67822c4c70dec1f812d31cc00748d50565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9b0b72fbe8b552d1266c310f6a6a67822c4c70dec1f812d31cc00748d50565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s5d7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:17Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.076706 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d87rr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a92ae69-5675-49d9-a567-e2476fe02ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://207d152294f9996ed7294d858842da5eef891a6d8a027473dc3d29a7101aadab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffwm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d87rr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:17Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.096669 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d40160e-075f-44b7-9d9d-3291e7774f27\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7079548fe9529583a1c129f57838c24cf06d1cb55a97c05c63a05d0cc2e7ec9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7762237003a5a256ff20a16a4869462e2e0dd1426aa391f7ec74ed9b1f572d4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1129 04:39:02.908489 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1129 04:39:02.908637 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 04:39:02.909677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703886220/tls.crt::/tmp/serving-cert-1703886220/tls.key\\\\\\\"\\\\nI1129 04:39:03.077876 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 04:39:03.080368 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 04:39:03.080390 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 04:39:03.080424 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 04:39:03.080432 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 04:39:03.086120 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 04:39:03.086240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086412 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 04:39:03.086502 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 04:39:03.086563 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 04:39:03.086622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1129 04:39:03.086141 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1129 04:39:03.086836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:38:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:17Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.110425 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://396c1c8cbd638c7dc4d8eebded6a17449d9abdc049161578581f1d8c3e31d9a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:17Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.122470 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:17Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.146640 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd007c61-4689-4179-95ad-44a9724e8e7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28499112c33332c23be39a11cdf7fc4e1946d04524a3f7c534087179a6620121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dac1c084955d874741d36709fa848c2c475a2d7108d4378b677bb13a88184105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72802532a47b381ead2507e1ae31c49852f9b1d332f3c08fb1e3a972a128eead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c81d94f098efec48c1ff8e172e476bf806ccec5777974eea9c0f3850e5fd40e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3d7f1ba1a408c54f0470fca40e307b70ba208dc7eeece599b684820e3f56c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d91c4f461a1912117cb8ed3a8087c5efaaf0a67455b721bfbeaad1c7dda8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea607ff09f683e7ce5d058b6b45b53cf355a787bed536da93bc6d52dac1334e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d135e42b8886f1f118abf8f4ce4cf2875ac4dc6c148bfc0677cef6b3651562c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fht8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:17Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.158370 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://deb9f370792598608cf2e0c0fe9975eaf8b95dc94b12eef8c55112df4aa9c998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:17Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.164477 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.164543 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.164599 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.164625 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.164642 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:17Z","lastTransitionTime":"2025-11-29T04:39:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.168424 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:17Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.179599 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee17b4a9a8999e5f76ceaf69dbd13acf80ed0e95e72f3e1428ce38c5d5982788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://382256e0fa1a4c26610227b9ae96288ccbfbbd97aa3cbd5fd4f18cb4228a234b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:17Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.193085 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt2dl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6c9236-5f7d-46c5-b080-4667620d874e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b6c827d0c57a87808fef301661fce28c50359e85c9e47947c50945f4f869de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt2dl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:17Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.208917 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s5d7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0af3d308-36b9-4622-8544-635f7b24d687\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0088b60069e830e9c36075b09d1ff135da79e94fe6cd271d0e93415af50557e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0088b60069e830e9c36075b09d1ff135da79e94fe6cd271d0e93415af50557e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9b0b72fbe8b552d1266c310f6a6a67822c4c70dec1f812d31cc00748d50565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9b0b72fbe8b552d1266c310f6a6a67822c4c70dec1f812d31cc00748d50565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa3b85ae7c47440f0f61436e1cb1f3f2096e0e4011434876f9d17e913ca24df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa3b85ae7c47440f0f61436e1cb1f3f2096e0e4011434876f9d17e913ca24df7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s5d7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:17Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.218203 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d87rr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a92ae69-5675-49d9-a567-e2476fe02ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://207d152294f9996ed7294d858842da5eef891a6d8a027473dc3d29a7101aadab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffwm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d87rr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:17Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.238147 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d40160e-075f-44b7-9d9d-3291e7774f27\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7079548fe9529583a1c129f57838c24cf06d1cb55a97c05c63a05d0cc2e7ec9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7762237003a5a256ff20a16a4869462e2e0dd1426aa391f7ec74ed9b1f572d4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1129 04:39:02.908489 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1129 04:39:02.908637 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 04:39:02.909677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703886220/tls.crt::/tmp/serving-cert-1703886220/tls.key\\\\\\\"\\\\nI1129 04:39:03.077876 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 04:39:03.080368 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 04:39:03.080390 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 04:39:03.080424 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 04:39:03.080432 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 04:39:03.086120 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 04:39:03.086240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086412 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 04:39:03.086502 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 04:39:03.086563 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 04:39:03.086622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1129 04:39:03.086141 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1129 04:39:03.086836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:38:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:17Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.252406 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://396c1c8cbd638c7dc4d8eebded6a17449d9abdc049161578581f1d8c3e31d9a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:17Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.264736 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:17Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.268611 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.268685 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.268701 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.268729 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.268754 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:17Z","lastTransitionTime":"2025-11-29T04:39:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.287765 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd007c61-4689-4179-95ad-44a9724e8e7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28499112c33332c23be39a11cdf7fc4e1946d04524a3f7c534087179a6620121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dac1c084955d874741d36709fa848c2c475a2d7108d4378b677bb13a88184105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72802532a47b381ead2507e1ae31c49852f9b1d332f3c08fb1e3a972a128eead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c81d94f098efec48c1ff8e172e476bf806ccec5777974eea9c0f3850e5fd40e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3d7f1ba1a408c54f0470fca40e307b70ba208dc7eeece599b684820e3f56c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d91c4f461a1912117cb8ed3a8087c5efaaf0a67455b721bfbeaad1c7dda8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea607ff09f683e7ce5d058b6b45b53cf355a787bed536da93bc6d52dac1334e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d135e42b8886f1f118abf8f4ce4cf2875ac4dc6c148bfc0677cef6b3651562c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fht8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:17Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.303859 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22ed28a-3748-4eb9-b3a7-c02aa38f7e94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33c0540ccb097b15f65369ff537aef4ae621cc828ba314eb61f0a0037608cadf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c241d29297a73dce12bd0bc2863c56475eb35b8fe8c93991ba0ef8695e38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f42a55d797c6c94af3a5eac1bf08aa38a82359b7ddd0ba32dcc797218ceee6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebed9ec415bb433935da36df8014f9087236cea9788b2048cbff8714f8ad0df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:17Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.315241 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:17Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.327464 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7lb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55156232-4c80-45a7-aad5-a5fa2793e4ed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6e5473c80737a5b7b00e5874062e0485ed7a4d74632202090a9b12f331953b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d2dtw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7lb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:17Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.345037 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0847696a7b7926047dc62d75739b3b6a7a04cd58efd1c1c2c992d5ae8d20ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b2ec8f0d384066ba90eb42c1a0a925ade3722b2515b0592d6f2a8f5e533a0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-26t88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:17Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.372050 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.372298 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.372437 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.372615 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.372761 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:17Z","lastTransitionTime":"2025-11-29T04:39:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.475667 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.475715 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.475728 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.475749 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.475761 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:17Z","lastTransitionTime":"2025-11-29T04:39:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.578666 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.578962 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.579032 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.579115 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.579186 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:17Z","lastTransitionTime":"2025-11-29T04:39:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.658556 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 04:39:17 crc kubenswrapper[4799]: E1129 04:39:17.658887 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.658970 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.659122 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:39:17 crc kubenswrapper[4799]: E1129 04:39:17.659282 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 04:39:17 crc kubenswrapper[4799]: E1129 04:39:17.659123 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.682092 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.682326 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.682403 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.682477 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.682540 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:17Z","lastTransitionTime":"2025-11-29T04:39:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.784615 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.784667 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.784682 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.784702 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.784716 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:17Z","lastTransitionTime":"2025-11-29T04:39:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.877964 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-s5d7s" event={"ID":"0af3d308-36b9-4622-8544-635f7b24d687","Type":"ContainerStarted","Data":"e23c8830e316099ef7b0e0044145a128658456871418c8c53bcc71210be290df"} Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.878091 4799 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.887234 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.887292 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.887305 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.887329 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.887346 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:17Z","lastTransitionTime":"2025-11-29T04:39:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.897414 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee17b4a9a8999e5f76ceaf69dbd13acf80ed0e95e72f3e1428ce38c5d5982788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://382256e0fa1a4c26610227b9ae96288ccbfbbd97aa3cbd5fd4f18cb4228a234b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:17Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.911778 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt2dl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6c9236-5f7d-46c5-b080-4667620d874e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b6c827d0c57a87808fef301661fce28c50359e85c9e47947c50945f4f869de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt2dl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:17Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.926067 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s5d7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0af3d308-36b9-4622-8544-635f7b24d687\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23c8830e316099ef7b0e0044145a128658456871418c8c53bcc71210be290df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0088b60069e830e9c36075b09d1ff135da79e94fe6cd271d0e93415af50557e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0088b60069e830e9c36075b09d1ff135da79e94fe6cd271d0e93415af50557e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9b0b72fbe8b552d1266c310f6a6a67822c4c70dec1f812d31cc00748d50565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9b0b72fbe8b552d1266c310f6a6a67822c4c70dec1f812d31cc00748d50565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa3b85ae7c47440f0f61436e1cb1f3f2096e0e4011434876f9d17e913ca24df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa3b85ae7c47440f0f61436e1cb1f3f2096e0e4011434876f9d17e913ca24df7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s5d7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:17Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.936294 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d87rr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a92ae69-5675-49d9-a567-e2476fe02ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://207d152294f9996ed7294d858842da5eef891a6d8a027473dc3d29a7101aadab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffwm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d87rr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:17Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.950897 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://deb9f370792598608cf2e0c0fe9975eaf8b95dc94b12eef8c55112df4aa9c998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:17Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.965133 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:17Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.989957 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d40160e-075f-44b7-9d9d-3291e7774f27\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7079548fe9529583a1c129f57838c24cf06d1cb55a97c05c63a05d0cc2e7ec9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7762237003a5a256ff20a16a4869462e2e0dd1426aa391f7ec74ed9b1f572d4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1129 04:39:02.908489 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1129 04:39:02.908637 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 04:39:02.909677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703886220/tls.crt::/tmp/serving-cert-1703886220/tls.key\\\\\\\"\\\\nI1129 04:39:03.077876 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 04:39:03.080368 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 04:39:03.080390 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 04:39:03.080424 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 04:39:03.080432 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 04:39:03.086120 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 04:39:03.086240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086412 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 04:39:03.086502 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 04:39:03.086563 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 04:39:03.086622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1129 04:39:03.086141 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1129 04:39:03.086836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:38:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:17Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.990725 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.990757 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.990768 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.990805 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.990820 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:17Z","lastTransitionTime":"2025-11-29T04:39:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.994675 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.994727 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.994740 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.994759 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:17 crc kubenswrapper[4799]: I1129 04:39:17.994772 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:17Z","lastTransitionTime":"2025-11-29T04:39:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.004689 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://396c1c8cbd638c7dc4d8eebded6a17449d9abdc049161578581f1d8c3e31d9a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:18Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:18 crc kubenswrapper[4799]: E1129 04:39:18.006601 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6f4e4672-da32-440c-8c4f-6e7f4942f55e\\\",\\\"systemUUID\\\":\\\"882c4d8e-57b7-4312-8499-40d1f85d56c9\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:18Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.009565 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.009625 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.009643 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.010076 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.010140 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:18Z","lastTransitionTime":"2025-11-29T04:39:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.017138 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:18Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:18 crc kubenswrapper[4799]: E1129 04:39:18.027259 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6f4e4672-da32-440c-8c4f-6e7f4942f55e\\\",\\\"systemUUID\\\":\\\"882c4d8e-57b7-4312-8499-40d1f85d56c9\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:18Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.032514 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.032559 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.032582 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.032603 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.032615 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:18Z","lastTransitionTime":"2025-11-29T04:39:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.035658 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd007c61-4689-4179-95ad-44a9724e8e7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28499112c33332c23be39a11cdf7fc4e1946d04524a3f7c534087179a6620121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dac1c084955d874741d36709fa848c2c475a2d7108d4378b677bb13a88184105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72802532a47b381ead2507e1ae31c49852f9b1d332f3c08fb1e3a972a128eead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c81d94f098efec48c1ff8e172e476bf806ccec5777974eea9c0f3850e5fd40e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3d7f1ba1a408c54f0470fca40e307b70ba208dc7eeece599b684820e3f56c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d91c4f461a1912117cb8ed3a8087c5efaaf0a67455b721bfbeaad1c7dda8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea607ff09f683e7ce5d058b6b45b53cf355a787bed536da93bc6d52dac1334e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d135e42b8886f1f118abf8f4ce4cf2875ac4dc6c148bfc0677cef6b3651562c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fht8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:18Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:18 crc kubenswrapper[4799]: E1129 04:39:18.043156 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6f4e4672-da32-440c-8c4f-6e7f4942f55e\\\",\\\"systemUUID\\\":\\\"882c4d8e-57b7-4312-8499-40d1f85d56c9\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:18Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.046144 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7lb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55156232-4c80-45a7-aad5-a5fa2793e4ed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6e5473c80737a5b7b00e5874062e0485ed7a4d74632202090a9b12f331953b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d2dtw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7lb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:18Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.046505 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.046540 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.046553 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.046572 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.046585 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:18Z","lastTransitionTime":"2025-11-29T04:39:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.057090 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0847696a7b7926047dc62d75739b3b6a7a04cd58efd1c1c2c992d5ae8d20ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b2ec8f0d384066ba90eb42c1a0a925ade3722b2515b0592d6f2a8f5e533a0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-26t88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:18Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:18 crc kubenswrapper[4799]: E1129 04:39:18.061283 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6f4e4672-da32-440c-8c4f-6e7f4942f55e\\\",\\\"systemUUID\\\":\\\"882c4d8e-57b7-4312-8499-40d1f85d56c9\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:18Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.065596 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.065623 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.065632 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.065646 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.065655 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:18Z","lastTransitionTime":"2025-11-29T04:39:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.070721 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22ed28a-3748-4eb9-b3a7-c02aa38f7e94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33c0540ccb097b15f65369ff537aef4ae621cc828ba314eb61f0a0037608cadf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c241d29297a73dce12bd0bc2863c56475eb35b8fe8c93991ba0ef8695e38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f42a55d797c6c94af3a5eac1bf08aa38a82359b7ddd0ba32dcc797218ceee6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebed9ec415bb433935da36df8014f9087236cea9788b2048cbff8714f8ad0df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:18Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:18 crc kubenswrapper[4799]: E1129 04:39:18.079104 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6f4e4672-da32-440c-8c4f-6e7f4942f55e\\\",\\\"systemUUID\\\":\\\"882c4d8e-57b7-4312-8499-40d1f85d56c9\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:18Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:18 crc kubenswrapper[4799]: E1129 04:39:18.079274 4799 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.085738 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:18Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.093750 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.093810 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.093826 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.093848 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.093863 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:18Z","lastTransitionTime":"2025-11-29T04:39:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.196098 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.196188 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.196201 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.196219 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.196232 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:18Z","lastTransitionTime":"2025-11-29T04:39:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.298212 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.298259 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.298274 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.298290 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.298302 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:18Z","lastTransitionTime":"2025-11-29T04:39:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.400521 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.400577 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.400591 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.400613 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.400626 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:18Z","lastTransitionTime":"2025-11-29T04:39:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.502383 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.502430 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.502442 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.502459 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.502470 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:18Z","lastTransitionTime":"2025-11-29T04:39:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.605732 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.605838 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.605863 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.605891 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.605911 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:18Z","lastTransitionTime":"2025-11-29T04:39:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.709005 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.709048 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.709060 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.709076 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.709087 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:18Z","lastTransitionTime":"2025-11-29T04:39:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.810648 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.810695 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.810713 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.810735 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.810752 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:18Z","lastTransitionTime":"2025-11-29T04:39:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.883307 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fht8l_bd007c61-4689-4179-95ad-44a9724e8e7e/ovnkube-controller/0.log" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.886898 4799 generic.go:334] "Generic (PLEG): container finished" podID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerID="ea607ff09f683e7ce5d058b6b45b53cf355a787bed536da93bc6d52dac1334e7" exitCode=1 Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.887899 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" event={"ID":"bd007c61-4689-4179-95ad-44a9724e8e7e","Type":"ContainerDied","Data":"ea607ff09f683e7ce5d058b6b45b53cf355a787bed536da93bc6d52dac1334e7"} Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.887930 4799 scope.go:117] "RemoveContainer" containerID="ea607ff09f683e7ce5d058b6b45b53cf355a787bed536da93bc6d52dac1334e7" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.910784 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee17b4a9a8999e5f76ceaf69dbd13acf80ed0e95e72f3e1428ce38c5d5982788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://382256e0fa1a4c26610227b9ae96288ccbfbbd97aa3cbd5fd4f18cb4228a234b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:18Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.913142 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.913184 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.913202 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.913233 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.913252 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:18Z","lastTransitionTime":"2025-11-29T04:39:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.927901 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt2dl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6c9236-5f7d-46c5-b080-4667620d874e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b6c827d0c57a87808fef301661fce28c50359e85c9e47947c50945f4f869de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt2dl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:18Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.943560 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s5d7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0af3d308-36b9-4622-8544-635f7b24d687\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23c8830e316099ef7b0e0044145a128658456871418c8c53bcc71210be290df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0088b60069e830e9c36075b09d1ff135da79e94fe6cd271d0e93415af50557e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0088b60069e830e9c36075b09d1ff135da79e94fe6cd271d0e93415af50557e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9b0b72fbe8b552d1266c310f6a6a67822c4c70dec1f812d31cc00748d50565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9b0b72fbe8b552d1266c310f6a6a67822c4c70dec1f812d31cc00748d50565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa3b85ae7c47440f0f61436e1cb1f3f2096e0e4011434876f9d17e913ca24df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa3b85ae7c47440f0f61436e1cb1f3f2096e0e4011434876f9d17e913ca24df7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s5d7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:18Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.957055 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d87rr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a92ae69-5675-49d9-a567-e2476fe02ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://207d152294f9996ed7294d858842da5eef891a6d8a027473dc3d29a7101aadab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffwm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d87rr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:18Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.976721 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://deb9f370792598608cf2e0c0fe9975eaf8b95dc94b12eef8c55112df4aa9c998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:18Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:18 crc kubenswrapper[4799]: I1129 04:39:18.995724 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:18Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.020982 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.021198 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.021218 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.021244 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.021266 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:19Z","lastTransitionTime":"2025-11-29T04:39:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.026532 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d40160e-075f-44b7-9d9d-3291e7774f27\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7079548fe9529583a1c129f57838c24cf06d1cb55a97c05c63a05d0cc2e7ec9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7762237003a5a256ff20a16a4869462e2e0dd1426aa391f7ec74ed9b1f572d4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1129 04:39:02.908489 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1129 04:39:02.908637 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 04:39:02.909677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703886220/tls.crt::/tmp/serving-cert-1703886220/tls.key\\\\\\\"\\\\nI1129 04:39:03.077876 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 04:39:03.080368 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 04:39:03.080390 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 04:39:03.080424 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 04:39:03.080432 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 04:39:03.086120 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 04:39:03.086240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086412 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 04:39:03.086502 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 04:39:03.086563 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 04:39:03.086622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1129 04:39:03.086141 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1129 04:39:03.086836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:38:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:19Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.048612 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://396c1c8cbd638c7dc4d8eebded6a17449d9abdc049161578581f1d8c3e31d9a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:19Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.071407 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:19Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.098724 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd007c61-4689-4179-95ad-44a9724e8e7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28499112c33332c23be39a11cdf7fc4e1946d04524a3f7c534087179a6620121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dac1c084955d874741d36709fa848c2c475a2d7108d4378b677bb13a88184105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72802532a47b381ead2507e1ae31c49852f9b1d332f3c08fb1e3a972a128eead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c81d94f098efec48c1ff8e172e476bf806ccec5777974eea9c0f3850e5fd40e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3d7f1ba1a408c54f0470fca40e307b70ba208dc7eeece599b684820e3f56c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d91c4f461a1912117cb8ed3a8087c5efaaf0a67455b721bfbeaad1c7dda8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea607ff09f683e7ce5d058b6b45b53cf355a787bed536da93bc6d52dac1334e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea607ff09f683e7ce5d058b6b45b53cf355a787bed536da93bc6d52dac1334e7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T04:39:18Z\\\",\\\"message\\\":\\\"r removal\\\\nI1129 04:39:18.659709 6119 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1129 04:39:18.659713 6119 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1129 04:39:18.659764 6119 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1129 04:39:18.659830 6119 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1129 04:39:18.659839 6119 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1129 04:39:18.659845 6119 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1129 04:39:18.659851 6119 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1129 04:39:18.659984 6119 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1129 04:39:18.660032 6119 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1129 04:39:18.660038 6119 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1129 04:39:18.660060 6119 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1129 04:39:18.660084 6119 factory.go:656] Stopping watch factory\\\\nI1129 04:39:18.660099 6119 ovnkube.go:599] Stopped ovnkube\\\\nI1129 04:39:18.660127 6119 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1129 04:39:18.660176 6119 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1129 04:39:18.660196 6119 handler.go:208] Removed *v1.Node eve\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d135e42b8886f1f118abf8f4ce4cf2875ac4dc6c148bfc0677cef6b3651562c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fht8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:19Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.118385 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7lb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55156232-4c80-45a7-aad5-a5fa2793e4ed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6e5473c80737a5b7b00e5874062e0485ed7a4d74632202090a9b12f331953b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d2dtw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7lb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:19Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.125235 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.125253 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.125262 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.125277 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.125289 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:19Z","lastTransitionTime":"2025-11-29T04:39:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.135760 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0847696a7b7926047dc62d75739b3b6a7a04cd58efd1c1c2c992d5ae8d20ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b2ec8f0d384066ba90eb42c1a0a925ade3722b2515b0592d6f2a8f5e533a0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-26t88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:19Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.156623 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22ed28a-3748-4eb9-b3a7-c02aa38f7e94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33c0540ccb097b15f65369ff537aef4ae621cc828ba314eb61f0a0037608cadf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c241d29297a73dce12bd0bc2863c56475eb35b8fe8c93991ba0ef8695e38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f42a55d797c6c94af3a5eac1bf08aa38a82359b7ddd0ba32dcc797218ceee6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebed9ec415bb433935da36df8014f9087236cea9788b2048cbff8714f8ad0df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:19Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.173351 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:19Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.227283 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.227339 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.227352 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.227369 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.227380 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:19Z","lastTransitionTime":"2025-11-29T04:39:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.329139 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.329172 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.329183 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.329199 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.329210 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:19Z","lastTransitionTime":"2025-11-29T04:39:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.432025 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.432067 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.432078 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.432094 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.432105 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:19Z","lastTransitionTime":"2025-11-29T04:39:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.534408 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.534463 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.534474 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.534492 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.534503 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:19Z","lastTransitionTime":"2025-11-29T04:39:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.565967 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.566110 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.566153 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:39:19 crc kubenswrapper[4799]: E1129 04:39:19.566246 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 04:39:35.566210262 +0000 UTC m=+51.209140712 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:39:19 crc kubenswrapper[4799]: E1129 04:39:19.566251 4799 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.566298 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.566340 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:39:19 crc kubenswrapper[4799]: E1129 04:39:19.566350 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-29 04:39:35.566336155 +0000 UTC m=+51.209266555 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 29 04:39:19 crc kubenswrapper[4799]: E1129 04:39:19.566347 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 29 04:39:19 crc kubenswrapper[4799]: E1129 04:39:19.566397 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 29 04:39:19 crc kubenswrapper[4799]: E1129 04:39:19.566414 4799 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 04:39:19 crc kubenswrapper[4799]: E1129 04:39:19.566449 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-29 04:39:35.566439527 +0000 UTC m=+51.209370167 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 04:39:19 crc kubenswrapper[4799]: E1129 04:39:19.566456 4799 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 29 04:39:19 crc kubenswrapper[4799]: E1129 04:39:19.566530 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-29 04:39:35.566509039 +0000 UTC m=+51.209439449 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 29 04:39:19 crc kubenswrapper[4799]: E1129 04:39:19.566535 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 29 04:39:19 crc kubenswrapper[4799]: E1129 04:39:19.566572 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 29 04:39:19 crc kubenswrapper[4799]: E1129 04:39:19.566590 4799 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 04:39:19 crc kubenswrapper[4799]: E1129 04:39:19.566651 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-29 04:39:35.566629992 +0000 UTC m=+51.209560572 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.637811 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.637865 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.637879 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.637907 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.637924 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:19Z","lastTransitionTime":"2025-11-29T04:39:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.658635 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.658746 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.658663 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 04:39:19 crc kubenswrapper[4799]: E1129 04:39:19.658866 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 04:39:19 crc kubenswrapper[4799]: E1129 04:39:19.658996 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 04:39:19 crc kubenswrapper[4799]: E1129 04:39:19.659131 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.741484 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.741537 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.741552 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.741574 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.741586 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:19Z","lastTransitionTime":"2025-11-29T04:39:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.843577 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.843609 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.843619 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.843633 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.843642 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:19Z","lastTransitionTime":"2025-11-29T04:39:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.895278 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fht8l_bd007c61-4689-4179-95ad-44a9724e8e7e/ovnkube-controller/0.log" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.897656 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" event={"ID":"bd007c61-4689-4179-95ad-44a9724e8e7e","Type":"ContainerStarted","Data":"1822ed94255ec78b74e15140fe5391fd667dd39afbda5fab6cdb350b924c833f"} Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.897721 4799 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.912096 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0847696a7b7926047dc62d75739b3b6a7a04cd58efd1c1c2c992d5ae8d20ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b2ec8f0d384066ba90eb42c1a0a925ade3722b2515b0592d6f2a8f5e533a0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-26t88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:19Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.923935 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22ed28a-3748-4eb9-b3a7-c02aa38f7e94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33c0540ccb097b15f65369ff537aef4ae621cc828ba314eb61f0a0037608cadf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c241d29297a73dce12bd0bc2863c56475eb35b8fe8c93991ba0ef8695e38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f42a55d797c6c94af3a5eac1bf08aa38a82359b7ddd0ba32dcc797218ceee6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebed9ec415bb433935da36df8014f9087236cea9788b2048cbff8714f8ad0df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:19Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.936684 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:19Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.945567 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.945592 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.945601 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.945614 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.945623 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:19Z","lastTransitionTime":"2025-11-29T04:39:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.948984 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7lb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55156232-4c80-45a7-aad5-a5fa2793e4ed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6e5473c80737a5b7b00e5874062e0485ed7a4d74632202090a9b12f331953b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d2dtw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7lb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:19Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.964331 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt2dl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6c9236-5f7d-46c5-b080-4667620d874e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b6c827d0c57a87808fef301661fce28c50359e85c9e47947c50945f4f869de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt2dl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:19Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.981678 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s5d7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0af3d308-36b9-4622-8544-635f7b24d687\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23c8830e316099ef7b0e0044145a128658456871418c8c53bcc71210be290df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0088b60069e830e9c36075b09d1ff135da79e94fe6cd271d0e93415af50557e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0088b60069e830e9c36075b09d1ff135da79e94fe6cd271d0e93415af50557e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9b0b72fbe8b552d1266c310f6a6a67822c4c70dec1f812d31cc00748d50565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9b0b72fbe8b552d1266c310f6a6a67822c4c70dec1f812d31cc00748d50565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa3b85ae7c47440f0f61436e1cb1f3f2096e0e4011434876f9d17e913ca24df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa3b85ae7c47440f0f61436e1cb1f3f2096e0e4011434876f9d17e913ca24df7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s5d7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:19Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:19 crc kubenswrapper[4799]: I1129 04:39:19.992604 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d87rr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a92ae69-5675-49d9-a567-e2476fe02ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://207d152294f9996ed7294d858842da5eef891a6d8a027473dc3d29a7101aadab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffwm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d87rr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:19Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.005530 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://deb9f370792598608cf2e0c0fe9975eaf8b95dc94b12eef8c55112df4aa9c998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:20Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.017979 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:20Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.034072 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee17b4a9a8999e5f76ceaf69dbd13acf80ed0e95e72f3e1428ce38c5d5982788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://382256e0fa1a4c26610227b9ae96288ccbfbbd97aa3cbd5fd4f18cb4228a234b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:20Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.048599 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.048633 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.048653 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.048671 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.048683 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:20Z","lastTransitionTime":"2025-11-29T04:39:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.051692 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d40160e-075f-44b7-9d9d-3291e7774f27\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7079548fe9529583a1c129f57838c24cf06d1cb55a97c05c63a05d0cc2e7ec9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7762237003a5a256ff20a16a4869462e2e0dd1426aa391f7ec74ed9b1f572d4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1129 04:39:02.908489 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1129 04:39:02.908637 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 04:39:02.909677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703886220/tls.crt::/tmp/serving-cert-1703886220/tls.key\\\\\\\"\\\\nI1129 04:39:03.077876 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 04:39:03.080368 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 04:39:03.080390 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 04:39:03.080424 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 04:39:03.080432 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 04:39:03.086120 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 04:39:03.086240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086412 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 04:39:03.086502 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 04:39:03.086563 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 04:39:03.086622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1129 04:39:03.086141 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1129 04:39:03.086836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:38:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:20Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.063996 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://396c1c8cbd638c7dc4d8eebded6a17449d9abdc049161578581f1d8c3e31d9a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:20Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.075093 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:20Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.091461 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd007c61-4689-4179-95ad-44a9724e8e7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28499112c33332c23be39a11cdf7fc4e1946d04524a3f7c534087179a6620121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dac1c084955d874741d36709fa848c2c475a2d7108d4378b677bb13a88184105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72802532a47b381ead2507e1ae31c49852f9b1d332f3c08fb1e3a972a128eead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c81d94f098efec48c1ff8e172e476bf806ccec5777974eea9c0f3850e5fd40e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3d7f1ba1a408c54f0470fca40e307b70ba208dc7eeece599b684820e3f56c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d91c4f461a1912117cb8ed3a8087c5efaaf0a67455b721bfbeaad1c7dda8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1822ed94255ec78b74e15140fe5391fd667dd39afbda5fab6cdb350b924c833f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea607ff09f683e7ce5d058b6b45b53cf355a787bed536da93bc6d52dac1334e7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T04:39:18Z\\\",\\\"message\\\":\\\"r removal\\\\nI1129 04:39:18.659709 6119 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1129 04:39:18.659713 6119 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1129 04:39:18.659764 6119 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1129 04:39:18.659830 6119 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1129 04:39:18.659839 6119 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1129 04:39:18.659845 6119 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1129 04:39:18.659851 6119 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1129 04:39:18.659984 6119 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1129 04:39:18.660032 6119 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1129 04:39:18.660038 6119 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1129 04:39:18.660060 6119 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1129 04:39:18.660084 6119 factory.go:656] Stopping watch factory\\\\nI1129 04:39:18.660099 6119 ovnkube.go:599] Stopped ovnkube\\\\nI1129 04:39:18.660127 6119 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1129 04:39:18.660176 6119 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1129 04:39:18.660196 6119 handler.go:208] Removed *v1.Node eve\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d135e42b8886f1f118abf8f4ce4cf2875ac4dc6c148bfc0677cef6b3651562c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fht8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:20Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.151659 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.151711 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.151721 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.151749 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.151760 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:20Z","lastTransitionTime":"2025-11-29T04:39:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.254368 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.254423 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.254442 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.254471 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.254495 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:20Z","lastTransitionTime":"2025-11-29T04:39:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.356663 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.356720 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.356740 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.356765 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.356783 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:20Z","lastTransitionTime":"2025-11-29T04:39:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.460016 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.460068 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.460078 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.460096 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.460109 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:20Z","lastTransitionTime":"2025-11-29T04:39:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.562550 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.562653 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.562677 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.562711 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.562734 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:20Z","lastTransitionTime":"2025-11-29T04:39:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.666290 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.666364 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.666379 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.666401 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.666418 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:20Z","lastTransitionTime":"2025-11-29T04:39:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.769463 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.769497 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.769508 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.769522 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.769532 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:20Z","lastTransitionTime":"2025-11-29T04:39:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.872751 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.872838 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.872857 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.872881 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.872897 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:20Z","lastTransitionTime":"2025-11-29T04:39:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.905569 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fht8l_bd007c61-4689-4179-95ad-44a9724e8e7e/ovnkube-controller/1.log" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.906247 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fht8l_bd007c61-4689-4179-95ad-44a9724e8e7e/ovnkube-controller/0.log" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.909844 4799 generic.go:334] "Generic (PLEG): container finished" podID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerID="1822ed94255ec78b74e15140fe5391fd667dd39afbda5fab6cdb350b924c833f" exitCode=1 Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.909932 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" event={"ID":"bd007c61-4689-4179-95ad-44a9724e8e7e","Type":"ContainerDied","Data":"1822ed94255ec78b74e15140fe5391fd667dd39afbda5fab6cdb350b924c833f"} Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.910041 4799 scope.go:117] "RemoveContainer" containerID="ea607ff09f683e7ce5d058b6b45b53cf355a787bed536da93bc6d52dac1334e7" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.910717 4799 scope.go:117] "RemoveContainer" containerID="1822ed94255ec78b74e15140fe5391fd667dd39afbda5fab6cdb350b924c833f" Nov 29 04:39:20 crc kubenswrapper[4799]: E1129 04:39:20.911000 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-fht8l_openshift-ovn-kubernetes(bd007c61-4689-4179-95ad-44a9724e8e7e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" podUID="bd007c61-4689-4179-95ad-44a9724e8e7e" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.934275 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22ed28a-3748-4eb9-b3a7-c02aa38f7e94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33c0540ccb097b15f65369ff537aef4ae621cc828ba314eb61f0a0037608cadf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c241d29297a73dce12bd0bc2863c56475eb35b8fe8c93991ba0ef8695e38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f42a55d797c6c94af3a5eac1bf08aa38a82359b7ddd0ba32dcc797218ceee6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebed9ec415bb433935da36df8014f9087236cea9788b2048cbff8714f8ad0df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:20Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.955554 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:20Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.976074 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.976127 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.976144 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.976172 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.976190 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:20Z","lastTransitionTime":"2025-11-29T04:39:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.977143 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7lb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55156232-4c80-45a7-aad5-a5fa2793e4ed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6e5473c80737a5b7b00e5874062e0485ed7a4d74632202090a9b12f331953b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d2dtw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7lb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:20Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:20 crc kubenswrapper[4799]: I1129 04:39:20.996460 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0847696a7b7926047dc62d75739b3b6a7a04cd58efd1c1c2c992d5ae8d20ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b2ec8f0d384066ba90eb42c1a0a925ade3722b2515b0592d6f2a8f5e533a0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-26t88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:20Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.022674 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s5d7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0af3d308-36b9-4622-8544-635f7b24d687\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23c8830e316099ef7b0e0044145a128658456871418c8c53bcc71210be290df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0088b60069e830e9c36075b09d1ff135da79e94fe6cd271d0e93415af50557e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0088b60069e830e9c36075b09d1ff135da79e94fe6cd271d0e93415af50557e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9b0b72fbe8b552d1266c310f6a6a67822c4c70dec1f812d31cc00748d50565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9b0b72fbe8b552d1266c310f6a6a67822c4c70dec1f812d31cc00748d50565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa3b85ae7c47440f0f61436e1cb1f3f2096e0e4011434876f9d17e913ca24df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa3b85ae7c47440f0f61436e1cb1f3f2096e0e4011434876f9d17e913ca24df7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s5d7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:21Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.041979 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d87rr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a92ae69-5675-49d9-a567-e2476fe02ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://207d152294f9996ed7294d858842da5eef891a6d8a027473dc3d29a7101aadab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffwm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d87rr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:21Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.067612 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://deb9f370792598608cf2e0c0fe9975eaf8b95dc94b12eef8c55112df4aa9c998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:21Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.079380 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.079433 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.079442 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.079461 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.079470 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:21Z","lastTransitionTime":"2025-11-29T04:39:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.103890 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:21Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.144579 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee17b4a9a8999e5f76ceaf69dbd13acf80ed0e95e72f3e1428ce38c5d5982788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://382256e0fa1a4c26610227b9ae96288ccbfbbd97aa3cbd5fd4f18cb4228a234b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:21Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.164479 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt2dl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6c9236-5f7d-46c5-b080-4667620d874e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b6c827d0c57a87808fef301661fce28c50359e85c9e47947c50945f4f869de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt2dl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:21Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.181772 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.181835 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.181847 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.181866 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.181877 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:21Z","lastTransitionTime":"2025-11-29T04:39:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.183611 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d40160e-075f-44b7-9d9d-3291e7774f27\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7079548fe9529583a1c129f57838c24cf06d1cb55a97c05c63a05d0cc2e7ec9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7762237003a5a256ff20a16a4869462e2e0dd1426aa391f7ec74ed9b1f572d4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1129 04:39:02.908489 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1129 04:39:02.908637 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 04:39:02.909677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703886220/tls.crt::/tmp/serving-cert-1703886220/tls.key\\\\\\\"\\\\nI1129 04:39:03.077876 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 04:39:03.080368 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 04:39:03.080390 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 04:39:03.080424 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 04:39:03.080432 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 04:39:03.086120 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 04:39:03.086240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086412 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 04:39:03.086502 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 04:39:03.086563 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 04:39:03.086622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1129 04:39:03.086141 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1129 04:39:03.086836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:38:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:21Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.198365 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://396c1c8cbd638c7dc4d8eebded6a17449d9abdc049161578581f1d8c3e31d9a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:21Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.214656 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:21Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.236223 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd007c61-4689-4179-95ad-44a9724e8e7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28499112c33332c23be39a11cdf7fc4e1946d04524a3f7c534087179a6620121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dac1c084955d874741d36709fa848c2c475a2d7108d4378b677bb13a88184105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72802532a47b381ead2507e1ae31c49852f9b1d332f3c08fb1e3a972a128eead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c81d94f098efec48c1ff8e172e476bf806ccec5777974eea9c0f3850e5fd40e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3d7f1ba1a408c54f0470fca40e307b70ba208dc7eeece599b684820e3f56c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d91c4f461a1912117cb8ed3a8087c5efaaf0a67455b721bfbeaad1c7dda8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1822ed94255ec78b74e15140fe5391fd667dd39afbda5fab6cdb350b924c833f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea607ff09f683e7ce5d058b6b45b53cf355a787bed536da93bc6d52dac1334e7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T04:39:18Z\\\",\\\"message\\\":\\\"r removal\\\\nI1129 04:39:18.659709 6119 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1129 04:39:18.659713 6119 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1129 04:39:18.659764 6119 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1129 04:39:18.659830 6119 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1129 04:39:18.659839 6119 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1129 04:39:18.659845 6119 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1129 04:39:18.659851 6119 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1129 04:39:18.659984 6119 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1129 04:39:18.660032 6119 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1129 04:39:18.660038 6119 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1129 04:39:18.660060 6119 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1129 04:39:18.660084 6119 factory.go:656] Stopping watch factory\\\\nI1129 04:39:18.660099 6119 ovnkube.go:599] Stopped ovnkube\\\\nI1129 04:39:18.660127 6119 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1129 04:39:18.660176 6119 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1129 04:39:18.660196 6119 handler.go:208] Removed *v1.Node eve\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1822ed94255ec78b74e15140fe5391fd667dd39afbda5fab6cdb350b924c833f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T04:39:19Z\\\",\\\"message\\\":\\\"890 6259 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1129 04:39:19.793922 6259 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1129 04:39:19.793930 6259 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1129 04:39:19.793939 6259 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1129 04:39:19.793959 6259 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1129 04:39:19.793967 6259 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1129 04:39:19.793998 6259 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1129 04:39:19.794016 6259 factory.go:656] Stopping watch factory\\\\nI1129 04:39:19.794027 6259 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1129 04:39:19.794034 6259 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1129 04:39:19.794040 6259 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1129 04:39:19.794048 6259 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1129 04:39:19.794053 6259 handler.go:208] Removed *v1.Node event handler 7\\\\nI1129 04:39:19.794060 6259 handler.go:208] Removed *v1.Node event handler 2\\\\nI1129 04:39:19.794109 6259 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d135e42b8886f1f118abf8f4ce4cf2875ac4dc6c148bfc0677cef6b3651562c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fht8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:21Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.284825 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.284852 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.284860 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.284875 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.284885 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:21Z","lastTransitionTime":"2025-11-29T04:39:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.386482 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.386513 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.386522 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.386534 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.386541 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:21Z","lastTransitionTime":"2025-11-29T04:39:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.488709 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.488745 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.488755 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.488767 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.488776 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:21Z","lastTransitionTime":"2025-11-29T04:39:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.590833 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.590859 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.590885 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.590899 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.590907 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:21Z","lastTransitionTime":"2025-11-29T04:39:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.658491 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 04:39:21 crc kubenswrapper[4799]: E1129 04:39:21.658605 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.658494 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.658492 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:39:21 crc kubenswrapper[4799]: E1129 04:39:21.658747 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 04:39:21 crc kubenswrapper[4799]: E1129 04:39:21.658694 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.692439 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.692489 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.692500 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.692512 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.692520 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:21Z","lastTransitionTime":"2025-11-29T04:39:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.794983 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.795034 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.795043 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.795055 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.795064 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:21Z","lastTransitionTime":"2025-11-29T04:39:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.897902 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.897943 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.897952 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.897967 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.897977 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:21Z","lastTransitionTime":"2025-11-29T04:39:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.914772 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fht8l_bd007c61-4689-4179-95ad-44a9724e8e7e/ovnkube-controller/1.log" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.919146 4799 scope.go:117] "RemoveContainer" containerID="1822ed94255ec78b74e15140fe5391fd667dd39afbda5fab6cdb350b924c833f" Nov 29 04:39:21 crc kubenswrapper[4799]: E1129 04:39:21.919423 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-fht8l_openshift-ovn-kubernetes(bd007c61-4689-4179-95ad-44a9724e8e7e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" podUID="bd007c61-4689-4179-95ad-44a9724e8e7e" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.934849 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22ed28a-3748-4eb9-b3a7-c02aa38f7e94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33c0540ccb097b15f65369ff537aef4ae621cc828ba314eb61f0a0037608cadf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c241d29297a73dce12bd0bc2863c56475eb35b8fe8c93991ba0ef8695e38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f42a55d797c6c94af3a5eac1bf08aa38a82359b7ddd0ba32dcc797218ceee6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebed9ec415bb433935da36df8014f9087236cea9788b2048cbff8714f8ad0df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:21Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.951876 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:21Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.958945 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.964596 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7lb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55156232-4c80-45a7-aad5-a5fa2793e4ed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6e5473c80737a5b7b00e5874062e0485ed7a4d74632202090a9b12f331953b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d2dtw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7lb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:21Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.980006 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0847696a7b7926047dc62d75739b3b6a7a04cd58efd1c1c2c992d5ae8d20ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b2ec8f0d384066ba90eb42c1a0a925ade3722b2515b0592d6f2a8f5e533a0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-26t88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:21Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:21 crc kubenswrapper[4799]: I1129 04:39:21.997060 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://deb9f370792598608cf2e0c0fe9975eaf8b95dc94b12eef8c55112df4aa9c998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:21Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.001988 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.002061 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.002080 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.002103 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.002120 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:22Z","lastTransitionTime":"2025-11-29T04:39:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.011235 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:22Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.026303 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee17b4a9a8999e5f76ceaf69dbd13acf80ed0e95e72f3e1428ce38c5d5982788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://382256e0fa1a4c26610227b9ae96288ccbfbbd97aa3cbd5fd4f18cb4228a234b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:22Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.042694 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt2dl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6c9236-5f7d-46c5-b080-4667620d874e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b6c827d0c57a87808fef301661fce28c50359e85c9e47947c50945f4f869de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt2dl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:22Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.058659 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s5d7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0af3d308-36b9-4622-8544-635f7b24d687\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23c8830e316099ef7b0e0044145a128658456871418c8c53bcc71210be290df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0088b60069e830e9c36075b09d1ff135da79e94fe6cd271d0e93415af50557e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0088b60069e830e9c36075b09d1ff135da79e94fe6cd271d0e93415af50557e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9b0b72fbe8b552d1266c310f6a6a67822c4c70dec1f812d31cc00748d50565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9b0b72fbe8b552d1266c310f6a6a67822c4c70dec1f812d31cc00748d50565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa3b85ae7c47440f0f61436e1cb1f3f2096e0e4011434876f9d17e913ca24df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa3b85ae7c47440f0f61436e1cb1f3f2096e0e4011434876f9d17e913ca24df7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s5d7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:22Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.073244 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d87rr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a92ae69-5675-49d9-a567-e2476fe02ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://207d152294f9996ed7294d858842da5eef891a6d8a027473dc3d29a7101aadab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffwm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d87rr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:22Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.090515 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d40160e-075f-44b7-9d9d-3291e7774f27\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7079548fe9529583a1c129f57838c24cf06d1cb55a97c05c63a05d0cc2e7ec9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7762237003a5a256ff20a16a4869462e2e0dd1426aa391f7ec74ed9b1f572d4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1129 04:39:02.908489 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1129 04:39:02.908637 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 04:39:02.909677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703886220/tls.crt::/tmp/serving-cert-1703886220/tls.key\\\\\\\"\\\\nI1129 04:39:03.077876 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 04:39:03.080368 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 04:39:03.080390 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 04:39:03.080424 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 04:39:03.080432 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 04:39:03.086120 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 04:39:03.086240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086412 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 04:39:03.086502 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 04:39:03.086563 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 04:39:03.086622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1129 04:39:03.086141 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1129 04:39:03.086836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:38:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:22Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.101267 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://396c1c8cbd638c7dc4d8eebded6a17449d9abdc049161578581f1d8c3e31d9a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:22Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.104424 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.104465 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.104475 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.104489 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.104498 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:22Z","lastTransitionTime":"2025-11-29T04:39:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.109043 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v79sj"] Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.109503 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v79sj" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.110991 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.113671 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.125195 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:22Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.146967 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd007c61-4689-4179-95ad-44a9724e8e7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28499112c33332c23be39a11cdf7fc4e1946d04524a3f7c534087179a6620121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dac1c084955d874741d36709fa848c2c475a2d7108d4378b677bb13a88184105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72802532a47b381ead2507e1ae31c49852f9b1d332f3c08fb1e3a972a128eead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c81d94f098efec48c1ff8e172e476bf806ccec5777974eea9c0f3850e5fd40e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3d7f1ba1a408c54f0470fca40e307b70ba208dc7eeece599b684820e3f56c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d91c4f461a1912117cb8ed3a8087c5efaaf0a67455b721bfbeaad1c7dda8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1822ed94255ec78b74e15140fe5391fd667dd39afbda5fab6cdb350b924c833f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1822ed94255ec78b74e15140fe5391fd667dd39afbda5fab6cdb350b924c833f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T04:39:19Z\\\",\\\"message\\\":\\\"890 6259 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1129 04:39:19.793922 6259 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1129 04:39:19.793930 6259 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1129 04:39:19.793939 6259 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1129 04:39:19.793959 6259 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1129 04:39:19.793967 6259 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1129 04:39:19.793998 6259 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1129 04:39:19.794016 6259 factory.go:656] Stopping watch factory\\\\nI1129 04:39:19.794027 6259 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1129 04:39:19.794034 6259 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1129 04:39:19.794040 6259 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1129 04:39:19.794048 6259 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1129 04:39:19.794053 6259 handler.go:208] Removed *v1.Node event handler 7\\\\nI1129 04:39:19.794060 6259 handler.go:208] Removed *v1.Node event handler 2\\\\nI1129 04:39:19.794109 6259 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-fht8l_openshift-ovn-kubernetes(bd007c61-4689-4179-95ad-44a9724e8e7e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d135e42b8886f1f118abf8f4ce4cf2875ac4dc6c148bfc0677cef6b3651562c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fht8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:22Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.162807 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:22Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.190741 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd007c61-4689-4179-95ad-44a9724e8e7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28499112c33332c23be39a11cdf7fc4e1946d04524a3f7c534087179a6620121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dac1c084955d874741d36709fa848c2c475a2d7108d4378b677bb13a88184105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72802532a47b381ead2507e1ae31c49852f9b1d332f3c08fb1e3a972a128eead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c81d94f098efec48c1ff8e172e476bf806ccec5777974eea9c0f3850e5fd40e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3d7f1ba1a408c54f0470fca40e307b70ba208dc7eeece599b684820e3f56c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d91c4f461a1912117cb8ed3a8087c5efaaf0a67455b721bfbeaad1c7dda8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1822ed94255ec78b74e15140fe5391fd667dd39afbda5fab6cdb350b924c833f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1822ed94255ec78b74e15140fe5391fd667dd39afbda5fab6cdb350b924c833f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T04:39:19Z\\\",\\\"message\\\":\\\"890 6259 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1129 04:39:19.793922 6259 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1129 04:39:19.793930 6259 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1129 04:39:19.793939 6259 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1129 04:39:19.793959 6259 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1129 04:39:19.793967 6259 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1129 04:39:19.793998 6259 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1129 04:39:19.794016 6259 factory.go:656] Stopping watch factory\\\\nI1129 04:39:19.794027 6259 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1129 04:39:19.794034 6259 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1129 04:39:19.794040 6259 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1129 04:39:19.794048 6259 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1129 04:39:19.794053 6259 handler.go:208] Removed *v1.Node event handler 7\\\\nI1129 04:39:19.794060 6259 handler.go:208] Removed *v1.Node event handler 2\\\\nI1129 04:39:19.794109 6259 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-fht8l_openshift-ovn-kubernetes(bd007c61-4689-4179-95ad-44a9724e8e7e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d135e42b8886f1f118abf8f4ce4cf2875ac4dc6c148bfc0677cef6b3651562c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fht8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:22Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.193146 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3a8cf268-e9fd-460a-b579-6411942a00a4-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-v79sj\" (UID: \"3a8cf268-e9fd-460a-b579-6411942a00a4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v79sj" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.193194 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3a8cf268-e9fd-460a-b579-6411942a00a4-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-v79sj\" (UID: \"3a8cf268-e9fd-460a-b579-6411942a00a4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v79sj" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.193362 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3a8cf268-e9fd-460a-b579-6411942a00a4-env-overrides\") pod \"ovnkube-control-plane-749d76644c-v79sj\" (UID: \"3a8cf268-e9fd-460a-b579-6411942a00a4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v79sj" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.193456 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bvxd\" (UniqueName: \"kubernetes.io/projected/3a8cf268-e9fd-460a-b579-6411942a00a4-kube-api-access-6bvxd\") pod \"ovnkube-control-plane-749d76644c-v79sj\" (UID: \"3a8cf268-e9fd-460a-b579-6411942a00a4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v79sj" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.207003 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.207048 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.207065 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.207088 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.207104 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:22Z","lastTransitionTime":"2025-11-29T04:39:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.207103 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7lb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55156232-4c80-45a7-aad5-a5fa2793e4ed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6e5473c80737a5b7b00e5874062e0485ed7a4d74632202090a9b12f331953b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d2dtw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7lb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:22Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.225433 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0847696a7b7926047dc62d75739b3b6a7a04cd58efd1c1c2c992d5ae8d20ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b2ec8f0d384066ba90eb42c1a0a925ade3722b2515b0592d6f2a8f5e533a0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-26t88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:22Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.238779 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22ed28a-3748-4eb9-b3a7-c02aa38f7e94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33c0540ccb097b15f65369ff537aef4ae621cc828ba314eb61f0a0037608cadf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c241d29297a73dce12bd0bc2863c56475eb35b8fe8c93991ba0ef8695e38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f42a55d797c6c94af3a5eac1bf08aa38a82359b7ddd0ba32dcc797218ceee6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebed9ec415bb433935da36df8014f9087236cea9788b2048cbff8714f8ad0df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:22Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.252266 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:22Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.269821 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee17b4a9a8999e5f76ceaf69dbd13acf80ed0e95e72f3e1428ce38c5d5982788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://382256e0fa1a4c26610227b9ae96288ccbfbbd97aa3cbd5fd4f18cb4228a234b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:22Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.283334 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt2dl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6c9236-5f7d-46c5-b080-4667620d874e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b6c827d0c57a87808fef301661fce28c50359e85c9e47947c50945f4f869de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt2dl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:22Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.294315 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3a8cf268-e9fd-460a-b579-6411942a00a4-env-overrides\") pod \"ovnkube-control-plane-749d76644c-v79sj\" (UID: \"3a8cf268-e9fd-460a-b579-6411942a00a4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v79sj" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.294369 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bvxd\" (UniqueName: \"kubernetes.io/projected/3a8cf268-e9fd-460a-b579-6411942a00a4-kube-api-access-6bvxd\") pod \"ovnkube-control-plane-749d76644c-v79sj\" (UID: \"3a8cf268-e9fd-460a-b579-6411942a00a4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v79sj" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.294408 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3a8cf268-e9fd-460a-b579-6411942a00a4-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-v79sj\" (UID: \"3a8cf268-e9fd-460a-b579-6411942a00a4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v79sj" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.294441 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3a8cf268-e9fd-460a-b579-6411942a00a4-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-v79sj\" (UID: \"3a8cf268-e9fd-460a-b579-6411942a00a4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v79sj" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.295184 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3a8cf268-e9fd-460a-b579-6411942a00a4-env-overrides\") pod \"ovnkube-control-plane-749d76644c-v79sj\" (UID: \"3a8cf268-e9fd-460a-b579-6411942a00a4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v79sj" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.295257 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3a8cf268-e9fd-460a-b579-6411942a00a4-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-v79sj\" (UID: \"3a8cf268-e9fd-460a-b579-6411942a00a4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v79sj" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.299188 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s5d7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0af3d308-36b9-4622-8544-635f7b24d687\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23c8830e316099ef7b0e0044145a128658456871418c8c53bcc71210be290df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0088b60069e830e9c36075b09d1ff135da79e94fe6cd271d0e93415af50557e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0088b60069e830e9c36075b09d1ff135da79e94fe6cd271d0e93415af50557e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9b0b72fbe8b552d1266c310f6a6a67822c4c70dec1f812d31cc00748d50565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9b0b72fbe8b552d1266c310f6a6a67822c4c70dec1f812d31cc00748d50565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa3b85ae7c47440f0f61436e1cb1f3f2096e0e4011434876f9d17e913ca24df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa3b85ae7c47440f0f61436e1cb1f3f2096e0e4011434876f9d17e913ca24df7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s5d7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:22Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.300629 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3a8cf268-e9fd-460a-b579-6411942a00a4-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-v79sj\" (UID: \"3a8cf268-e9fd-460a-b579-6411942a00a4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v79sj" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.309241 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.309302 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.309315 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.309331 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.309341 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:22Z","lastTransitionTime":"2025-11-29T04:39:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.310745 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d87rr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a92ae69-5675-49d9-a567-e2476fe02ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://207d152294f9996ed7294d858842da5eef891a6d8a027473dc3d29a7101aadab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffwm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d87rr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:22Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.315195 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bvxd\" (UniqueName: \"kubernetes.io/projected/3a8cf268-e9fd-460a-b579-6411942a00a4-kube-api-access-6bvxd\") pod \"ovnkube-control-plane-749d76644c-v79sj\" (UID: \"3a8cf268-e9fd-460a-b579-6411942a00a4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v79sj" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.328502 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://deb9f370792598608cf2e0c0fe9975eaf8b95dc94b12eef8c55112df4aa9c998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:22Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.341135 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:22Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.354312 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d40160e-075f-44b7-9d9d-3291e7774f27\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7079548fe9529583a1c129f57838c24cf06d1cb55a97c05c63a05d0cc2e7ec9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7762237003a5a256ff20a16a4869462e2e0dd1426aa391f7ec74ed9b1f572d4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1129 04:39:02.908489 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1129 04:39:02.908637 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 04:39:02.909677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703886220/tls.crt::/tmp/serving-cert-1703886220/tls.key\\\\\\\"\\\\nI1129 04:39:03.077876 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 04:39:03.080368 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 04:39:03.080390 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 04:39:03.080424 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 04:39:03.080432 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 04:39:03.086120 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 04:39:03.086240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086412 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 04:39:03.086502 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 04:39:03.086563 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 04:39:03.086622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1129 04:39:03.086141 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1129 04:39:03.086836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:38:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:22Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.367112 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://396c1c8cbd638c7dc4d8eebded6a17449d9abdc049161578581f1d8c3e31d9a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:22Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.379046 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v79sj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8cf268-e9fd-460a-b579-6411942a00a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v79sj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:22Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.412172 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.412204 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.412218 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.412247 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.412257 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:22Z","lastTransitionTime":"2025-11-29T04:39:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.424508 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v79sj" Nov 29 04:39:22 crc kubenswrapper[4799]: W1129 04:39:22.435355 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3a8cf268_e9fd_460a_b579_6411942a00a4.slice/crio-72556a7ba4db43e92f46c67efd4168c33520379e3da9791fbcf8f3faeda952f0 WatchSource:0}: Error finding container 72556a7ba4db43e92f46c67efd4168c33520379e3da9791fbcf8f3faeda952f0: Status 404 returned error can't find the container with id 72556a7ba4db43e92f46c67efd4168c33520379e3da9791fbcf8f3faeda952f0 Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.514318 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.514356 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.514366 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.514381 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.514395 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:22Z","lastTransitionTime":"2025-11-29T04:39:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.616465 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.616504 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.616516 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.616532 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.616543 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:22Z","lastTransitionTime":"2025-11-29T04:39:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.718463 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.718510 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.718526 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.718547 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.718564 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:22Z","lastTransitionTime":"2025-11-29T04:39:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.821027 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.821095 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.821122 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.821153 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.821179 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:22Z","lastTransitionTime":"2025-11-29T04:39:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.922281 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v79sj" event={"ID":"3a8cf268-e9fd-460a-b579-6411942a00a4","Type":"ContainerStarted","Data":"6c4e099a3149affd1f9f9c5cbaa0f2758d6ab92117c7352db8ede170153c6bc3"} Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.922344 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v79sj" event={"ID":"3a8cf268-e9fd-460a-b579-6411942a00a4","Type":"ContainerStarted","Data":"72556a7ba4db43e92f46c67efd4168c33520379e3da9791fbcf8f3faeda952f0"} Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.922395 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.922424 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.922433 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.922447 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:22 crc kubenswrapper[4799]: I1129 04:39:22.922457 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:22Z","lastTransitionTime":"2025-11-29T04:39:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.025599 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.025636 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.025647 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.025664 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.025676 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:23Z","lastTransitionTime":"2025-11-29T04:39:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.128149 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.128189 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.128199 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.128214 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.128223 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:23Z","lastTransitionTime":"2025-11-29T04:39:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.230131 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.230181 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.230194 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.230212 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.230223 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:23Z","lastTransitionTime":"2025-11-29T04:39:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.333414 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.333472 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.333596 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.333617 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.333629 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:23Z","lastTransitionTime":"2025-11-29T04:39:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.436597 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.436648 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.436659 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.436677 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.436688 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:23Z","lastTransitionTime":"2025-11-29T04:39:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.539098 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.539157 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.539175 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.539197 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.539213 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:23Z","lastTransitionTime":"2025-11-29T04:39:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.641375 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.641454 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.641478 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.641513 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.641537 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:23Z","lastTransitionTime":"2025-11-29T04:39:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.658023 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.658097 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.658041 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 04:39:23 crc kubenswrapper[4799]: E1129 04:39:23.658189 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 04:39:23 crc kubenswrapper[4799]: E1129 04:39:23.658346 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 04:39:23 crc kubenswrapper[4799]: E1129 04:39:23.658469 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.744554 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.744608 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.744625 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.744646 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.744660 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:23Z","lastTransitionTime":"2025-11-29T04:39:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.847441 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.847510 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.847527 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.847547 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.847562 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:23Z","lastTransitionTime":"2025-11-29T04:39:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.925778 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v79sj" event={"ID":"3a8cf268-e9fd-460a-b579-6411942a00a4","Type":"ContainerStarted","Data":"93f943aedb7ef6351aa6b2a46b3e84b6db2fae1145bd40f3b9aee6d897aa0d5c"} Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.946429 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://deb9f370792598608cf2e0c0fe9975eaf8b95dc94b12eef8c55112df4aa9c998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:23Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.950525 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.950589 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.950605 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.950630 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.950646 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:23Z","lastTransitionTime":"2025-11-29T04:39:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.963465 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:23Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.966995 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-qx9wn"] Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.967536 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qx9wn" Nov 29 04:39:23 crc kubenswrapper[4799]: E1129 04:39:23.967635 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qx9wn" podUID="ce8ce318-27e6-434d-96c8-591d250262fd" Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.978627 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee17b4a9a8999e5f76ceaf69dbd13acf80ed0e95e72f3e1428ce38c5d5982788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://382256e0fa1a4c26610227b9ae96288ccbfbbd97aa3cbd5fd4f18cb4228a234b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:23Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:23 crc kubenswrapper[4799]: I1129 04:39:23.994632 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt2dl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6c9236-5f7d-46c5-b080-4667620d874e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b6c827d0c57a87808fef301661fce28c50359e85c9e47947c50945f4f869de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt2dl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:23Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.010902 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s5d7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0af3d308-36b9-4622-8544-635f7b24d687\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23c8830e316099ef7b0e0044145a128658456871418c8c53bcc71210be290df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0088b60069e830e9c36075b09d1ff135da79e94fe6cd271d0e93415af50557e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0088b60069e830e9c36075b09d1ff135da79e94fe6cd271d0e93415af50557e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9b0b72fbe8b552d1266c310f6a6a67822c4c70dec1f812d31cc00748d50565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9b0b72fbe8b552d1266c310f6a6a67822c4c70dec1f812d31cc00748d50565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa3b85ae7c47440f0f61436e1cb1f3f2096e0e4011434876f9d17e913ca24df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa3b85ae7c47440f0f61436e1cb1f3f2096e0e4011434876f9d17e913ca24df7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s5d7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:24Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.022422 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d87rr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a92ae69-5675-49d9-a567-e2476fe02ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://207d152294f9996ed7294d858842da5eef891a6d8a027473dc3d29a7101aadab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffwm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d87rr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:24Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.034756 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://396c1c8cbd638c7dc4d8eebded6a17449d9abdc049161578581f1d8c3e31d9a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:24Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.048893 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v79sj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8cf268-e9fd-460a-b579-6411942a00a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c4e099a3149affd1f9f9c5cbaa0f2758d6ab92117c7352db8ede170153c6bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f943aedb7ef6351aa6b2a46b3e84b6db2fae1145bd40f3b9aee6d897aa0d5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v79sj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:24Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.052922 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.052970 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.052983 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.053000 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.053010 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:24Z","lastTransitionTime":"2025-11-29T04:39:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.064081 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d40160e-075f-44b7-9d9d-3291e7774f27\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7079548fe9529583a1c129f57838c24cf06d1cb55a97c05c63a05d0cc2e7ec9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7762237003a5a256ff20a16a4869462e2e0dd1426aa391f7ec74ed9b1f572d4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1129 04:39:02.908489 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1129 04:39:02.908637 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 04:39:02.909677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703886220/tls.crt::/tmp/serving-cert-1703886220/tls.key\\\\\\\"\\\\nI1129 04:39:03.077876 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 04:39:03.080368 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 04:39:03.080390 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 04:39:03.080424 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 04:39:03.080432 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 04:39:03.086120 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 04:39:03.086240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086412 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 04:39:03.086502 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 04:39:03.086563 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 04:39:03.086622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1129 04:39:03.086141 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1129 04:39:03.086836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:38:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:24Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.076382 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:24Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.102652 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd007c61-4689-4179-95ad-44a9724e8e7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28499112c33332c23be39a11cdf7fc4e1946d04524a3f7c534087179a6620121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dac1c084955d874741d36709fa848c2c475a2d7108d4378b677bb13a88184105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72802532a47b381ead2507e1ae31c49852f9b1d332f3c08fb1e3a972a128eead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c81d94f098efec48c1ff8e172e476bf806ccec5777974eea9c0f3850e5fd40e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3d7f1ba1a408c54f0470fca40e307b70ba208dc7eeece599b684820e3f56c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d91c4f461a1912117cb8ed3a8087c5efaaf0a67455b721bfbeaad1c7dda8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1822ed94255ec78b74e15140fe5391fd667dd39afbda5fab6cdb350b924c833f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1822ed94255ec78b74e15140fe5391fd667dd39afbda5fab6cdb350b924c833f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T04:39:19Z\\\",\\\"message\\\":\\\"890 6259 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1129 04:39:19.793922 6259 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1129 04:39:19.793930 6259 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1129 04:39:19.793939 6259 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1129 04:39:19.793959 6259 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1129 04:39:19.793967 6259 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1129 04:39:19.793998 6259 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1129 04:39:19.794016 6259 factory.go:656] Stopping watch factory\\\\nI1129 04:39:19.794027 6259 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1129 04:39:19.794034 6259 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1129 04:39:19.794040 6259 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1129 04:39:19.794048 6259 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1129 04:39:19.794053 6259 handler.go:208] Removed *v1.Node event handler 7\\\\nI1129 04:39:19.794060 6259 handler.go:208] Removed *v1.Node event handler 2\\\\nI1129 04:39:19.794109 6259 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-fht8l_openshift-ovn-kubernetes(bd007c61-4689-4179-95ad-44a9724e8e7e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d135e42b8886f1f118abf8f4ce4cf2875ac4dc6c148bfc0677cef6b3651562c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fht8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:24Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.114909 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c99mf\" (UniqueName: \"kubernetes.io/projected/ce8ce318-27e6-434d-96c8-591d250262fd-kube-api-access-c99mf\") pod \"network-metrics-daemon-qx9wn\" (UID: \"ce8ce318-27e6-434d-96c8-591d250262fd\") " pod="openshift-multus/network-metrics-daemon-qx9wn" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.114976 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ce8ce318-27e6-434d-96c8-591d250262fd-metrics-certs\") pod \"network-metrics-daemon-qx9wn\" (UID: \"ce8ce318-27e6-434d-96c8-591d250262fd\") " pod="openshift-multus/network-metrics-daemon-qx9wn" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.116415 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:24Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.127310 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7lb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55156232-4c80-45a7-aad5-a5fa2793e4ed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6e5473c80737a5b7b00e5874062e0485ed7a4d74632202090a9b12f331953b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d2dtw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7lb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:24Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.141368 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0847696a7b7926047dc62d75739b3b6a7a04cd58efd1c1c2c992d5ae8d20ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b2ec8f0d384066ba90eb42c1a0a925ade3722b2515b0592d6f2a8f5e533a0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-26t88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:24Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.156539 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.156577 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.156586 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.156602 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.156612 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:24Z","lastTransitionTime":"2025-11-29T04:39:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.158562 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22ed28a-3748-4eb9-b3a7-c02aa38f7e94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33c0540ccb097b15f65369ff537aef4ae621cc828ba314eb61f0a0037608cadf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c241d29297a73dce12bd0bc2863c56475eb35b8fe8c93991ba0ef8695e38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f42a55d797c6c94af3a5eac1bf08aa38a82359b7ddd0ba32dcc797218ceee6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebed9ec415bb433935da36df8014f9087236cea9788b2048cbff8714f8ad0df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:24Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.175525 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://deb9f370792598608cf2e0c0fe9975eaf8b95dc94b12eef8c55112df4aa9c998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:24Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.187190 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:24Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.201447 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee17b4a9a8999e5f76ceaf69dbd13acf80ed0e95e72f3e1428ce38c5d5982788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://382256e0fa1a4c26610227b9ae96288ccbfbbd97aa3cbd5fd4f18cb4228a234b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:24Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.214774 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt2dl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6c9236-5f7d-46c5-b080-4667620d874e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b6c827d0c57a87808fef301661fce28c50359e85c9e47947c50945f4f869de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt2dl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:24Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.216282 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ce8ce318-27e6-434d-96c8-591d250262fd-metrics-certs\") pod \"network-metrics-daemon-qx9wn\" (UID: \"ce8ce318-27e6-434d-96c8-591d250262fd\") " pod="openshift-multus/network-metrics-daemon-qx9wn" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.216415 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c99mf\" (UniqueName: \"kubernetes.io/projected/ce8ce318-27e6-434d-96c8-591d250262fd-kube-api-access-c99mf\") pod \"network-metrics-daemon-qx9wn\" (UID: \"ce8ce318-27e6-434d-96c8-591d250262fd\") " pod="openshift-multus/network-metrics-daemon-qx9wn" Nov 29 04:39:24 crc kubenswrapper[4799]: E1129 04:39:24.216499 4799 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 29 04:39:24 crc kubenswrapper[4799]: E1129 04:39:24.216594 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ce8ce318-27e6-434d-96c8-591d250262fd-metrics-certs podName:ce8ce318-27e6-434d-96c8-591d250262fd nodeName:}" failed. No retries permitted until 2025-11-29 04:39:24.716573032 +0000 UTC m=+40.359503442 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ce8ce318-27e6-434d-96c8-591d250262fd-metrics-certs") pod "network-metrics-daemon-qx9wn" (UID: "ce8ce318-27e6-434d-96c8-591d250262fd") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.234306 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c99mf\" (UniqueName: \"kubernetes.io/projected/ce8ce318-27e6-434d-96c8-591d250262fd-kube-api-access-c99mf\") pod \"network-metrics-daemon-qx9wn\" (UID: \"ce8ce318-27e6-434d-96c8-591d250262fd\") " pod="openshift-multus/network-metrics-daemon-qx9wn" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.237731 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s5d7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0af3d308-36b9-4622-8544-635f7b24d687\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23c8830e316099ef7b0e0044145a128658456871418c8c53bcc71210be290df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0088b60069e830e9c36075b09d1ff135da79e94fe6cd271d0e93415af50557e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0088b60069e830e9c36075b09d1ff135da79e94fe6cd271d0e93415af50557e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9b0b72fbe8b552d1266c310f6a6a67822c4c70dec1f812d31cc00748d50565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9b0b72fbe8b552d1266c310f6a6a67822c4c70dec1f812d31cc00748d50565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa3b85ae7c47440f0f61436e1cb1f3f2096e0e4011434876f9d17e913ca24df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa3b85ae7c47440f0f61436e1cb1f3f2096e0e4011434876f9d17e913ca24df7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s5d7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:24Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.252749 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d87rr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a92ae69-5675-49d9-a567-e2476fe02ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://207d152294f9996ed7294d858842da5eef891a6d8a027473dc3d29a7101aadab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffwm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d87rr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:24Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.260046 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.260104 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.260123 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.260151 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.260185 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:24Z","lastTransitionTime":"2025-11-29T04:39:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.269146 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://396c1c8cbd638c7dc4d8eebded6a17449d9abdc049161578581f1d8c3e31d9a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:24Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.286531 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v79sj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8cf268-e9fd-460a-b579-6411942a00a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c4e099a3149affd1f9f9c5cbaa0f2758d6ab92117c7352db8ede170153c6bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f943aedb7ef6351aa6b2a46b3e84b6db2fae1145bd40f3b9aee6d897aa0d5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v79sj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:24Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.304479 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d40160e-075f-44b7-9d9d-3291e7774f27\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7079548fe9529583a1c129f57838c24cf06d1cb55a97c05c63a05d0cc2e7ec9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7762237003a5a256ff20a16a4869462e2e0dd1426aa391f7ec74ed9b1f572d4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1129 04:39:02.908489 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1129 04:39:02.908637 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 04:39:02.909677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703886220/tls.crt::/tmp/serving-cert-1703886220/tls.key\\\\\\\"\\\\nI1129 04:39:03.077876 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 04:39:03.080368 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 04:39:03.080390 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 04:39:03.080424 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 04:39:03.080432 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 04:39:03.086120 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 04:39:03.086240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086412 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 04:39:03.086502 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 04:39:03.086563 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 04:39:03.086622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1129 04:39:03.086141 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1129 04:39:03.086836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:38:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:24Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.320621 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:24Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.344120 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd007c61-4689-4179-95ad-44a9724e8e7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28499112c33332c23be39a11cdf7fc4e1946d04524a3f7c534087179a6620121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dac1c084955d874741d36709fa848c2c475a2d7108d4378b677bb13a88184105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72802532a47b381ead2507e1ae31c49852f9b1d332f3c08fb1e3a972a128eead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c81d94f098efec48c1ff8e172e476bf806ccec5777974eea9c0f3850e5fd40e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3d7f1ba1a408c54f0470fca40e307b70ba208dc7eeece599b684820e3f56c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d91c4f461a1912117cb8ed3a8087c5efaaf0a67455b721bfbeaad1c7dda8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1822ed94255ec78b74e15140fe5391fd667dd39afbda5fab6cdb350b924c833f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1822ed94255ec78b74e15140fe5391fd667dd39afbda5fab6cdb350b924c833f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T04:39:19Z\\\",\\\"message\\\":\\\"890 6259 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1129 04:39:19.793922 6259 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1129 04:39:19.793930 6259 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1129 04:39:19.793939 6259 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1129 04:39:19.793959 6259 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1129 04:39:19.793967 6259 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1129 04:39:19.793998 6259 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1129 04:39:19.794016 6259 factory.go:656] Stopping watch factory\\\\nI1129 04:39:19.794027 6259 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1129 04:39:19.794034 6259 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1129 04:39:19.794040 6259 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1129 04:39:19.794048 6259 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1129 04:39:19.794053 6259 handler.go:208] Removed *v1.Node event handler 7\\\\nI1129 04:39:19.794060 6259 handler.go:208] Removed *v1.Node event handler 2\\\\nI1129 04:39:19.794109 6259 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-fht8l_openshift-ovn-kubernetes(bd007c61-4689-4179-95ad-44a9724e8e7e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d135e42b8886f1f118abf8f4ce4cf2875ac4dc6c148bfc0677cef6b3651562c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fht8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:24Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.360680 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:24Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.362840 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.362907 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.362947 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.362972 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.362986 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:24Z","lastTransitionTime":"2025-11-29T04:39:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.376112 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7lb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55156232-4c80-45a7-aad5-a5fa2793e4ed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6e5473c80737a5b7b00e5874062e0485ed7a4d74632202090a9b12f331953b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d2dtw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7lb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:24Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.390859 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0847696a7b7926047dc62d75739b3b6a7a04cd58efd1c1c2c992d5ae8d20ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b2ec8f0d384066ba90eb42c1a0a925ade3722b2515b0592d6f2a8f5e533a0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-26t88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:24Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.406862 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-qx9wn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce8ce318-27e6-434d-96c8-591d250262fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c99mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c99mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-qx9wn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:24Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.421505 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22ed28a-3748-4eb9-b3a7-c02aa38f7e94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33c0540ccb097b15f65369ff537aef4ae621cc828ba314eb61f0a0037608cadf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c241d29297a73dce12bd0bc2863c56475eb35b8fe8c93991ba0ef8695e38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f42a55d797c6c94af3a5eac1bf08aa38a82359b7ddd0ba32dcc797218ceee6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebed9ec415bb433935da36df8014f9087236cea9788b2048cbff8714f8ad0df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:24Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.466321 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.466375 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.466392 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.466413 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.466427 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:24Z","lastTransitionTime":"2025-11-29T04:39:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.570171 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.570228 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.570238 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.570255 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.570267 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:24Z","lastTransitionTime":"2025-11-29T04:39:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.672475 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.672556 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.672584 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.672619 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.672646 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:24Z","lastTransitionTime":"2025-11-29T04:39:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.681645 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22ed28a-3748-4eb9-b3a7-c02aa38f7e94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33c0540ccb097b15f65369ff537aef4ae621cc828ba314eb61f0a0037608cadf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c241d29297a73dce12bd0bc2863c56475eb35b8fe8c93991ba0ef8695e38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f42a55d797c6c94af3a5eac1bf08aa38a82359b7ddd0ba32dcc797218ceee6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebed9ec415bb433935da36df8014f9087236cea9788b2048cbff8714f8ad0df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:24Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.699340 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:24Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.714038 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7lb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55156232-4c80-45a7-aad5-a5fa2793e4ed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6e5473c80737a5b7b00e5874062e0485ed7a4d74632202090a9b12f331953b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d2dtw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7lb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:24Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.723284 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ce8ce318-27e6-434d-96c8-591d250262fd-metrics-certs\") pod \"network-metrics-daemon-qx9wn\" (UID: \"ce8ce318-27e6-434d-96c8-591d250262fd\") " pod="openshift-multus/network-metrics-daemon-qx9wn" Nov 29 04:39:24 crc kubenswrapper[4799]: E1129 04:39:24.723494 4799 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 29 04:39:24 crc kubenswrapper[4799]: E1129 04:39:24.723607 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ce8ce318-27e6-434d-96c8-591d250262fd-metrics-certs podName:ce8ce318-27e6-434d-96c8-591d250262fd nodeName:}" failed. No retries permitted until 2025-11-29 04:39:25.72357661 +0000 UTC m=+41.366507050 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ce8ce318-27e6-434d-96c8-591d250262fd-metrics-certs") pod "network-metrics-daemon-qx9wn" (UID: "ce8ce318-27e6-434d-96c8-591d250262fd") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.734868 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0847696a7b7926047dc62d75739b3b6a7a04cd58efd1c1c2c992d5ae8d20ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b2ec8f0d384066ba90eb42c1a0a925ade3722b2515b0592d6f2a8f5e533a0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-26t88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:24Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.749663 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-qx9wn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce8ce318-27e6-434d-96c8-591d250262fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c99mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c99mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-qx9wn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:24Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.764717 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://deb9f370792598608cf2e0c0fe9975eaf8b95dc94b12eef8c55112df4aa9c998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:24Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.775656 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.775709 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.775727 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.775747 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.775759 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:24Z","lastTransitionTime":"2025-11-29T04:39:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.784575 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:24Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.801037 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee17b4a9a8999e5f76ceaf69dbd13acf80ed0e95e72f3e1428ce38c5d5982788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://382256e0fa1a4c26610227b9ae96288ccbfbbd97aa3cbd5fd4f18cb4228a234b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:24Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.823766 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt2dl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6c9236-5f7d-46c5-b080-4667620d874e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b6c827d0c57a87808fef301661fce28c50359e85c9e47947c50945f4f869de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt2dl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:24Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.844156 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s5d7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0af3d308-36b9-4622-8544-635f7b24d687\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23c8830e316099ef7b0e0044145a128658456871418c8c53bcc71210be290df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0088b60069e830e9c36075b09d1ff135da79e94fe6cd271d0e93415af50557e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0088b60069e830e9c36075b09d1ff135da79e94fe6cd271d0e93415af50557e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9b0b72fbe8b552d1266c310f6a6a67822c4c70dec1f812d31cc00748d50565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9b0b72fbe8b552d1266c310f6a6a67822c4c70dec1f812d31cc00748d50565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa3b85ae7c47440f0f61436e1cb1f3f2096e0e4011434876f9d17e913ca24df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa3b85ae7c47440f0f61436e1cb1f3f2096e0e4011434876f9d17e913ca24df7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s5d7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:24Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.857948 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d87rr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a92ae69-5675-49d9-a567-e2476fe02ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://207d152294f9996ed7294d858842da5eef891a6d8a027473dc3d29a7101aadab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffwm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d87rr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:24Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.878298 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d40160e-075f-44b7-9d9d-3291e7774f27\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7079548fe9529583a1c129f57838c24cf06d1cb55a97c05c63a05d0cc2e7ec9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7762237003a5a256ff20a16a4869462e2e0dd1426aa391f7ec74ed9b1f572d4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1129 04:39:02.908489 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1129 04:39:02.908637 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 04:39:02.909677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703886220/tls.crt::/tmp/serving-cert-1703886220/tls.key\\\\\\\"\\\\nI1129 04:39:03.077876 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 04:39:03.080368 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 04:39:03.080390 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 04:39:03.080424 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 04:39:03.080432 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 04:39:03.086120 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 04:39:03.086240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086412 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 04:39:03.086502 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 04:39:03.086563 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 04:39:03.086622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1129 04:39:03.086141 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1129 04:39:03.086836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:38:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:24Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.879266 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.879337 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.879360 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.879394 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.879416 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:24Z","lastTransitionTime":"2025-11-29T04:39:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.891831 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://396c1c8cbd638c7dc4d8eebded6a17449d9abdc049161578581f1d8c3e31d9a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:24Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.908848 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v79sj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8cf268-e9fd-460a-b579-6411942a00a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c4e099a3149affd1f9f9c5cbaa0f2758d6ab92117c7352db8ede170153c6bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f943aedb7ef6351aa6b2a46b3e84b6db2fae1145bd40f3b9aee6d897aa0d5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v79sj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:24Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.924903 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:24Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.945975 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd007c61-4689-4179-95ad-44a9724e8e7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28499112c33332c23be39a11cdf7fc4e1946d04524a3f7c534087179a6620121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dac1c084955d874741d36709fa848c2c475a2d7108d4378b677bb13a88184105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72802532a47b381ead2507e1ae31c49852f9b1d332f3c08fb1e3a972a128eead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c81d94f098efec48c1ff8e172e476bf806ccec5777974eea9c0f3850e5fd40e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3d7f1ba1a408c54f0470fca40e307b70ba208dc7eeece599b684820e3f56c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d91c4f461a1912117cb8ed3a8087c5efaaf0a67455b721bfbeaad1c7dda8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1822ed94255ec78b74e15140fe5391fd667dd39afbda5fab6cdb350b924c833f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1822ed94255ec78b74e15140fe5391fd667dd39afbda5fab6cdb350b924c833f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T04:39:19Z\\\",\\\"message\\\":\\\"890 6259 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1129 04:39:19.793922 6259 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1129 04:39:19.793930 6259 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1129 04:39:19.793939 6259 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1129 04:39:19.793959 6259 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1129 04:39:19.793967 6259 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1129 04:39:19.793998 6259 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1129 04:39:19.794016 6259 factory.go:656] Stopping watch factory\\\\nI1129 04:39:19.794027 6259 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1129 04:39:19.794034 6259 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1129 04:39:19.794040 6259 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1129 04:39:19.794048 6259 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1129 04:39:19.794053 6259 handler.go:208] Removed *v1.Node event handler 7\\\\nI1129 04:39:19.794060 6259 handler.go:208] Removed *v1.Node event handler 2\\\\nI1129 04:39:19.794109 6259 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-fht8l_openshift-ovn-kubernetes(bd007c61-4689-4179-95ad-44a9724e8e7e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d135e42b8886f1f118abf8f4ce4cf2875ac4dc6c148bfc0677cef6b3651562c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fht8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:24Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.981525 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.981564 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.981574 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.981589 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:24 crc kubenswrapper[4799]: I1129 04:39:24.981599 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:24Z","lastTransitionTime":"2025-11-29T04:39:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:25 crc kubenswrapper[4799]: I1129 04:39:25.083707 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:25 crc kubenswrapper[4799]: I1129 04:39:25.083744 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:25 crc kubenswrapper[4799]: I1129 04:39:25.083756 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:25 crc kubenswrapper[4799]: I1129 04:39:25.083772 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:25 crc kubenswrapper[4799]: I1129 04:39:25.083801 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:25Z","lastTransitionTime":"2025-11-29T04:39:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:25 crc kubenswrapper[4799]: I1129 04:39:25.186167 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:25 crc kubenswrapper[4799]: I1129 04:39:25.186202 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:25 crc kubenswrapper[4799]: I1129 04:39:25.186212 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:25 crc kubenswrapper[4799]: I1129 04:39:25.186225 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:25 crc kubenswrapper[4799]: I1129 04:39:25.186235 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:25Z","lastTransitionTime":"2025-11-29T04:39:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:25 crc kubenswrapper[4799]: I1129 04:39:25.287901 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:25 crc kubenswrapper[4799]: I1129 04:39:25.287964 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:25 crc kubenswrapper[4799]: I1129 04:39:25.287975 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:25 crc kubenswrapper[4799]: I1129 04:39:25.287994 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:25 crc kubenswrapper[4799]: I1129 04:39:25.288005 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:25Z","lastTransitionTime":"2025-11-29T04:39:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:25 crc kubenswrapper[4799]: I1129 04:39:25.390271 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:25 crc kubenswrapper[4799]: I1129 04:39:25.390350 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:25 crc kubenswrapper[4799]: I1129 04:39:25.390366 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:25 crc kubenswrapper[4799]: I1129 04:39:25.390387 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:25 crc kubenswrapper[4799]: I1129 04:39:25.390403 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:25Z","lastTransitionTime":"2025-11-29T04:39:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:25 crc kubenswrapper[4799]: I1129 04:39:25.493103 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:25 crc kubenswrapper[4799]: I1129 04:39:25.493145 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:25 crc kubenswrapper[4799]: I1129 04:39:25.493158 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:25 crc kubenswrapper[4799]: I1129 04:39:25.493175 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:25 crc kubenswrapper[4799]: I1129 04:39:25.493186 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:25Z","lastTransitionTime":"2025-11-29T04:39:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:25 crc kubenswrapper[4799]: I1129 04:39:25.595850 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:25 crc kubenswrapper[4799]: I1129 04:39:25.595904 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:25 crc kubenswrapper[4799]: I1129 04:39:25.595914 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:25 crc kubenswrapper[4799]: I1129 04:39:25.595929 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:25 crc kubenswrapper[4799]: I1129 04:39:25.595938 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:25Z","lastTransitionTime":"2025-11-29T04:39:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:25 crc kubenswrapper[4799]: I1129 04:39:25.658625 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qx9wn" Nov 29 04:39:25 crc kubenswrapper[4799]: I1129 04:39:25.658653 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 04:39:25 crc kubenswrapper[4799]: I1129 04:39:25.658671 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 04:39:25 crc kubenswrapper[4799]: I1129 04:39:25.658670 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:39:25 crc kubenswrapper[4799]: E1129 04:39:25.658946 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qx9wn" podUID="ce8ce318-27e6-434d-96c8-591d250262fd" Nov 29 04:39:25 crc kubenswrapper[4799]: E1129 04:39:25.659068 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 04:39:25 crc kubenswrapper[4799]: E1129 04:39:25.659168 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 04:39:25 crc kubenswrapper[4799]: E1129 04:39:25.659235 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 04:39:25 crc kubenswrapper[4799]: I1129 04:39:25.698475 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:25 crc kubenswrapper[4799]: I1129 04:39:25.698508 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:25 crc kubenswrapper[4799]: I1129 04:39:25.698517 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:25 crc kubenswrapper[4799]: I1129 04:39:25.698533 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:25 crc kubenswrapper[4799]: I1129 04:39:25.698546 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:25Z","lastTransitionTime":"2025-11-29T04:39:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:25 crc kubenswrapper[4799]: I1129 04:39:25.733104 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ce8ce318-27e6-434d-96c8-591d250262fd-metrics-certs\") pod \"network-metrics-daemon-qx9wn\" (UID: \"ce8ce318-27e6-434d-96c8-591d250262fd\") " pod="openshift-multus/network-metrics-daemon-qx9wn" Nov 29 04:39:25 crc kubenswrapper[4799]: E1129 04:39:25.733300 4799 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 29 04:39:25 crc kubenswrapper[4799]: E1129 04:39:25.733388 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ce8ce318-27e6-434d-96c8-591d250262fd-metrics-certs podName:ce8ce318-27e6-434d-96c8-591d250262fd nodeName:}" failed. No retries permitted until 2025-11-29 04:39:27.733368045 +0000 UTC m=+43.376298445 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ce8ce318-27e6-434d-96c8-591d250262fd-metrics-certs") pod "network-metrics-daemon-qx9wn" (UID: "ce8ce318-27e6-434d-96c8-591d250262fd") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 29 04:39:25 crc kubenswrapper[4799]: I1129 04:39:25.801854 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:25 crc kubenswrapper[4799]: I1129 04:39:25.801927 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:25 crc kubenswrapper[4799]: I1129 04:39:25.801939 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:25 crc kubenswrapper[4799]: I1129 04:39:25.801960 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:25 crc kubenswrapper[4799]: I1129 04:39:25.801973 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:25Z","lastTransitionTime":"2025-11-29T04:39:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:25 crc kubenswrapper[4799]: I1129 04:39:25.904235 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:25 crc kubenswrapper[4799]: I1129 04:39:25.904277 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:25 crc kubenswrapper[4799]: I1129 04:39:25.904288 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:25 crc kubenswrapper[4799]: I1129 04:39:25.904305 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:25 crc kubenswrapper[4799]: I1129 04:39:25.904316 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:25Z","lastTransitionTime":"2025-11-29T04:39:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:26 crc kubenswrapper[4799]: I1129 04:39:26.007015 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:26 crc kubenswrapper[4799]: I1129 04:39:26.007068 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:26 crc kubenswrapper[4799]: I1129 04:39:26.007084 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:26 crc kubenswrapper[4799]: I1129 04:39:26.007107 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:26 crc kubenswrapper[4799]: I1129 04:39:26.007124 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:26Z","lastTransitionTime":"2025-11-29T04:39:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:26 crc kubenswrapper[4799]: I1129 04:39:26.109203 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:26 crc kubenswrapper[4799]: I1129 04:39:26.109273 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:26 crc kubenswrapper[4799]: I1129 04:39:26.109287 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:26 crc kubenswrapper[4799]: I1129 04:39:26.109303 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:26 crc kubenswrapper[4799]: I1129 04:39:26.109315 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:26Z","lastTransitionTime":"2025-11-29T04:39:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:26 crc kubenswrapper[4799]: I1129 04:39:26.211939 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:26 crc kubenswrapper[4799]: I1129 04:39:26.212040 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:26 crc kubenswrapper[4799]: I1129 04:39:26.212064 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:26 crc kubenswrapper[4799]: I1129 04:39:26.212100 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:26 crc kubenswrapper[4799]: I1129 04:39:26.212121 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:26Z","lastTransitionTime":"2025-11-29T04:39:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:26 crc kubenswrapper[4799]: I1129 04:39:26.317050 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:26 crc kubenswrapper[4799]: I1129 04:39:26.317137 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:26 crc kubenswrapper[4799]: I1129 04:39:26.317169 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:26 crc kubenswrapper[4799]: I1129 04:39:26.317204 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:26 crc kubenswrapper[4799]: I1129 04:39:26.317228 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:26Z","lastTransitionTime":"2025-11-29T04:39:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:26 crc kubenswrapper[4799]: I1129 04:39:26.422228 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:26 crc kubenswrapper[4799]: I1129 04:39:26.422314 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:26 crc kubenswrapper[4799]: I1129 04:39:26.422339 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:26 crc kubenswrapper[4799]: I1129 04:39:26.422377 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:26 crc kubenswrapper[4799]: I1129 04:39:26.422401 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:26Z","lastTransitionTime":"2025-11-29T04:39:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:26 crc kubenswrapper[4799]: I1129 04:39:26.526154 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:26 crc kubenswrapper[4799]: I1129 04:39:26.526226 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:26 crc kubenswrapper[4799]: I1129 04:39:26.526249 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:26 crc kubenswrapper[4799]: I1129 04:39:26.526280 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:26 crc kubenswrapper[4799]: I1129 04:39:26.526301 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:26Z","lastTransitionTime":"2025-11-29T04:39:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:26 crc kubenswrapper[4799]: I1129 04:39:26.629263 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:26 crc kubenswrapper[4799]: I1129 04:39:26.629317 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:26 crc kubenswrapper[4799]: I1129 04:39:26.629333 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:26 crc kubenswrapper[4799]: I1129 04:39:26.629354 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:26 crc kubenswrapper[4799]: I1129 04:39:26.629368 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:26Z","lastTransitionTime":"2025-11-29T04:39:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:26 crc kubenswrapper[4799]: I1129 04:39:26.733446 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:26 crc kubenswrapper[4799]: I1129 04:39:26.733532 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:26 crc kubenswrapper[4799]: I1129 04:39:26.733553 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:26 crc kubenswrapper[4799]: I1129 04:39:26.733589 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:26 crc kubenswrapper[4799]: I1129 04:39:26.733613 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:26Z","lastTransitionTime":"2025-11-29T04:39:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:26 crc kubenswrapper[4799]: I1129 04:39:26.837184 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:26 crc kubenswrapper[4799]: I1129 04:39:26.837249 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:26 crc kubenswrapper[4799]: I1129 04:39:26.837267 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:26 crc kubenswrapper[4799]: I1129 04:39:26.837292 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:26 crc kubenswrapper[4799]: I1129 04:39:26.837310 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:26Z","lastTransitionTime":"2025-11-29T04:39:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:26 crc kubenswrapper[4799]: I1129 04:39:26.940160 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:26 crc kubenswrapper[4799]: I1129 04:39:26.940214 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:26 crc kubenswrapper[4799]: I1129 04:39:26.940226 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:26 crc kubenswrapper[4799]: I1129 04:39:26.940249 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:26 crc kubenswrapper[4799]: I1129 04:39:26.940263 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:26Z","lastTransitionTime":"2025-11-29T04:39:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.043372 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.043433 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.043447 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.043471 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.043486 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:27Z","lastTransitionTime":"2025-11-29T04:39:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.147263 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.147342 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.147380 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.147405 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.147418 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:27Z","lastTransitionTime":"2025-11-29T04:39:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.251362 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.251407 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.251420 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.251442 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.251456 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:27Z","lastTransitionTime":"2025-11-29T04:39:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.354561 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.354635 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.354654 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.354686 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.354707 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:27Z","lastTransitionTime":"2025-11-29T04:39:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.458150 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.458227 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.458247 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.458281 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.458344 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:27Z","lastTransitionTime":"2025-11-29T04:39:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.561867 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.561956 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.561981 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.562017 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.562042 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:27Z","lastTransitionTime":"2025-11-29T04:39:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.658085 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.658177 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qx9wn" Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.658295 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 04:39:27 crc kubenswrapper[4799]: E1129 04:39:27.658334 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.658090 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:39:27 crc kubenswrapper[4799]: E1129 04:39:27.658474 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qx9wn" podUID="ce8ce318-27e6-434d-96c8-591d250262fd" Nov 29 04:39:27 crc kubenswrapper[4799]: E1129 04:39:27.658744 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 04:39:27 crc kubenswrapper[4799]: E1129 04:39:27.658950 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.665548 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.665607 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.665625 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.665648 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.665667 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:27Z","lastTransitionTime":"2025-11-29T04:39:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.759057 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ce8ce318-27e6-434d-96c8-591d250262fd-metrics-certs\") pod \"network-metrics-daemon-qx9wn\" (UID: \"ce8ce318-27e6-434d-96c8-591d250262fd\") " pod="openshift-multus/network-metrics-daemon-qx9wn" Nov 29 04:39:27 crc kubenswrapper[4799]: E1129 04:39:27.759351 4799 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 29 04:39:27 crc kubenswrapper[4799]: E1129 04:39:27.759534 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ce8ce318-27e6-434d-96c8-591d250262fd-metrics-certs podName:ce8ce318-27e6-434d-96c8-591d250262fd nodeName:}" failed. No retries permitted until 2025-11-29 04:39:31.759488645 +0000 UTC m=+47.402419085 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ce8ce318-27e6-434d-96c8-591d250262fd-metrics-certs") pod "network-metrics-daemon-qx9wn" (UID: "ce8ce318-27e6-434d-96c8-591d250262fd") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.769994 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.770061 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.770078 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.770107 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.770125 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:27Z","lastTransitionTime":"2025-11-29T04:39:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.873457 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.873539 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.873560 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.873597 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.873618 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:27Z","lastTransitionTime":"2025-11-29T04:39:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.977530 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.977613 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.977633 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.977666 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:27 crc kubenswrapper[4799]: I1129 04:39:27.977695 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:27Z","lastTransitionTime":"2025-11-29T04:39:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.081266 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.081347 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.081369 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.081403 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.081426 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:28Z","lastTransitionTime":"2025-11-29T04:39:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.185195 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.185273 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.185293 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.185321 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.185341 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:28Z","lastTransitionTime":"2025-11-29T04:39:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.289180 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.289248 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.289265 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.289297 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.289315 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:28Z","lastTransitionTime":"2025-11-29T04:39:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.373513 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.373606 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.373627 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.373657 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.373678 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:28Z","lastTransitionTime":"2025-11-29T04:39:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:28 crc kubenswrapper[4799]: E1129 04:39:28.402745 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6f4e4672-da32-440c-8c4f-6e7f4942f55e\\\",\\\"systemUUID\\\":\\\"882c4d8e-57b7-4312-8499-40d1f85d56c9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:28Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.410378 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.410438 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.410466 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.410502 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.410531 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:28Z","lastTransitionTime":"2025-11-29T04:39:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:28 crc kubenswrapper[4799]: E1129 04:39:28.433524 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6f4e4672-da32-440c-8c4f-6e7f4942f55e\\\",\\\"systemUUID\\\":\\\"882c4d8e-57b7-4312-8499-40d1f85d56c9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:28Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.439079 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.439133 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.439152 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.439177 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.439197 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:28Z","lastTransitionTime":"2025-11-29T04:39:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:28 crc kubenswrapper[4799]: E1129 04:39:28.462022 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6f4e4672-da32-440c-8c4f-6e7f4942f55e\\\",\\\"systemUUID\\\":\\\"882c4d8e-57b7-4312-8499-40d1f85d56c9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:28Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.467670 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.467853 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.467886 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.467959 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.467989 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:28Z","lastTransitionTime":"2025-11-29T04:39:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:28 crc kubenswrapper[4799]: E1129 04:39:28.490920 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6f4e4672-da32-440c-8c4f-6e7f4942f55e\\\",\\\"systemUUID\\\":\\\"882c4d8e-57b7-4312-8499-40d1f85d56c9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:28Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.496994 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.497062 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.497087 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.497120 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.497146 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:28Z","lastTransitionTime":"2025-11-29T04:39:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:28 crc kubenswrapper[4799]: E1129 04:39:28.518364 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6f4e4672-da32-440c-8c4f-6e7f4942f55e\\\",\\\"systemUUID\\\":\\\"882c4d8e-57b7-4312-8499-40d1f85d56c9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:28Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:28 crc kubenswrapper[4799]: E1129 04:39:28.518596 4799 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.520513 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.520663 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.520741 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.520783 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.520836 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:28Z","lastTransitionTime":"2025-11-29T04:39:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.624675 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.624740 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.624758 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.624821 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.624846 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:28Z","lastTransitionTime":"2025-11-29T04:39:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.728354 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.728436 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.728463 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.728495 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.728523 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:28Z","lastTransitionTime":"2025-11-29T04:39:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.832051 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.832157 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.832179 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.832208 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.832230 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:28Z","lastTransitionTime":"2025-11-29T04:39:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.935987 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.936186 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.936216 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.936252 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:28 crc kubenswrapper[4799]: I1129 04:39:28.936276 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:28Z","lastTransitionTime":"2025-11-29T04:39:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.040041 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.040114 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.040134 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.040162 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.040185 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:29Z","lastTransitionTime":"2025-11-29T04:39:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.142396 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.142436 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.142448 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.142465 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.142477 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:29Z","lastTransitionTime":"2025-11-29T04:39:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.246102 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.246150 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.246163 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.246185 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.246207 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:29Z","lastTransitionTime":"2025-11-29T04:39:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.348603 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.348662 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.348676 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.348697 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.348712 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:29Z","lastTransitionTime":"2025-11-29T04:39:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.451562 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.451637 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.451647 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.451664 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.451676 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:29Z","lastTransitionTime":"2025-11-29T04:39:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.554831 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.554892 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.554905 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.554925 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.554956 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:29Z","lastTransitionTime":"2025-11-29T04:39:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.657198 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.657239 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.657251 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.657269 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.657281 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:29Z","lastTransitionTime":"2025-11-29T04:39:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.658252 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.658266 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.658266 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.658350 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qx9wn" Nov 29 04:39:29 crc kubenswrapper[4799]: E1129 04:39:29.658367 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 04:39:29 crc kubenswrapper[4799]: E1129 04:39:29.658452 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 04:39:29 crc kubenswrapper[4799]: E1129 04:39:29.658521 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qx9wn" podUID="ce8ce318-27e6-434d-96c8-591d250262fd" Nov 29 04:39:29 crc kubenswrapper[4799]: E1129 04:39:29.658768 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.760316 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.760385 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.760403 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.760428 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.760445 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:29Z","lastTransitionTime":"2025-11-29T04:39:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.863751 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.863864 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.863889 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.863920 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.863944 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:29Z","lastTransitionTime":"2025-11-29T04:39:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.966772 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.966851 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.966865 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.966883 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:29 crc kubenswrapper[4799]: I1129 04:39:29.966905 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:29Z","lastTransitionTime":"2025-11-29T04:39:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:30 crc kubenswrapper[4799]: I1129 04:39:30.070435 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:30 crc kubenswrapper[4799]: I1129 04:39:30.070499 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:30 crc kubenswrapper[4799]: I1129 04:39:30.070522 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:30 crc kubenswrapper[4799]: I1129 04:39:30.070553 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:30 crc kubenswrapper[4799]: I1129 04:39:30.070575 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:30Z","lastTransitionTime":"2025-11-29T04:39:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:30 crc kubenswrapper[4799]: I1129 04:39:30.174220 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:30 crc kubenswrapper[4799]: I1129 04:39:30.174277 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:30 crc kubenswrapper[4799]: I1129 04:39:30.174292 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:30 crc kubenswrapper[4799]: I1129 04:39:30.174314 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:30 crc kubenswrapper[4799]: I1129 04:39:30.174329 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:30Z","lastTransitionTime":"2025-11-29T04:39:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:30 crc kubenswrapper[4799]: I1129 04:39:30.277384 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:30 crc kubenswrapper[4799]: I1129 04:39:30.277438 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:30 crc kubenswrapper[4799]: I1129 04:39:30.277457 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:30 crc kubenswrapper[4799]: I1129 04:39:30.277481 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:30 crc kubenswrapper[4799]: I1129 04:39:30.277496 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:30Z","lastTransitionTime":"2025-11-29T04:39:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:30 crc kubenswrapper[4799]: I1129 04:39:30.380834 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:30 crc kubenswrapper[4799]: I1129 04:39:30.380881 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:30 crc kubenswrapper[4799]: I1129 04:39:30.380897 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:30 crc kubenswrapper[4799]: I1129 04:39:30.380922 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:30 crc kubenswrapper[4799]: I1129 04:39:30.380938 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:30Z","lastTransitionTime":"2025-11-29T04:39:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:30 crc kubenswrapper[4799]: I1129 04:39:30.484277 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:30 crc kubenswrapper[4799]: I1129 04:39:30.484341 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:30 crc kubenswrapper[4799]: I1129 04:39:30.484359 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:30 crc kubenswrapper[4799]: I1129 04:39:30.484391 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:30 crc kubenswrapper[4799]: I1129 04:39:30.484451 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:30Z","lastTransitionTime":"2025-11-29T04:39:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:30 crc kubenswrapper[4799]: I1129 04:39:30.586524 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:30 crc kubenswrapper[4799]: I1129 04:39:30.586599 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:30 crc kubenswrapper[4799]: I1129 04:39:30.586619 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:30 crc kubenswrapper[4799]: I1129 04:39:30.586651 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:30 crc kubenswrapper[4799]: I1129 04:39:30.586671 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:30Z","lastTransitionTime":"2025-11-29T04:39:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:30 crc kubenswrapper[4799]: I1129 04:39:30.689691 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:30 crc kubenswrapper[4799]: I1129 04:39:30.689771 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:30 crc kubenswrapper[4799]: I1129 04:39:30.689859 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:30 crc kubenswrapper[4799]: I1129 04:39:30.689899 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:30 crc kubenswrapper[4799]: I1129 04:39:30.689924 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:30Z","lastTransitionTime":"2025-11-29T04:39:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:30 crc kubenswrapper[4799]: I1129 04:39:30.794379 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:30 crc kubenswrapper[4799]: I1129 04:39:30.794433 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:30 crc kubenswrapper[4799]: I1129 04:39:30.794444 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:30 crc kubenswrapper[4799]: I1129 04:39:30.794464 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:30 crc kubenswrapper[4799]: I1129 04:39:30.794479 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:30Z","lastTransitionTime":"2025-11-29T04:39:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:30 crc kubenswrapper[4799]: I1129 04:39:30.897936 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:30 crc kubenswrapper[4799]: I1129 04:39:30.898004 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:30 crc kubenswrapper[4799]: I1129 04:39:30.898020 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:30 crc kubenswrapper[4799]: I1129 04:39:30.898044 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:30 crc kubenswrapper[4799]: I1129 04:39:30.898062 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:30Z","lastTransitionTime":"2025-11-29T04:39:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.001112 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.001185 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.001205 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.001232 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.001251 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:31Z","lastTransitionTime":"2025-11-29T04:39:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.104516 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.104598 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.104622 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.104652 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.104680 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:31Z","lastTransitionTime":"2025-11-29T04:39:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.208222 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.208305 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.208329 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.208360 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.208382 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:31Z","lastTransitionTime":"2025-11-29T04:39:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.312180 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.312278 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.312307 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.312342 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.312366 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:31Z","lastTransitionTime":"2025-11-29T04:39:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.415751 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.415890 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.415921 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.415964 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.416004 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:31Z","lastTransitionTime":"2025-11-29T04:39:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.519329 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.519397 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.519416 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.519444 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.519463 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:31Z","lastTransitionTime":"2025-11-29T04:39:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.622934 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.623016 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.623044 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.623082 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.623108 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:31Z","lastTransitionTime":"2025-11-29T04:39:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.658906 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qx9wn" Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.658939 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.659005 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.659076 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:39:31 crc kubenswrapper[4799]: E1129 04:39:31.659272 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qx9wn" podUID="ce8ce318-27e6-434d-96c8-591d250262fd" Nov 29 04:39:31 crc kubenswrapper[4799]: E1129 04:39:31.659627 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 04:39:31 crc kubenswrapper[4799]: E1129 04:39:31.659834 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 04:39:31 crc kubenswrapper[4799]: E1129 04:39:31.660089 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.726476 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.726549 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.726571 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.726602 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.726626 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:31Z","lastTransitionTime":"2025-11-29T04:39:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.807176 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ce8ce318-27e6-434d-96c8-591d250262fd-metrics-certs\") pod \"network-metrics-daemon-qx9wn\" (UID: \"ce8ce318-27e6-434d-96c8-591d250262fd\") " pod="openshift-multus/network-metrics-daemon-qx9wn" Nov 29 04:39:31 crc kubenswrapper[4799]: E1129 04:39:31.807399 4799 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 29 04:39:31 crc kubenswrapper[4799]: E1129 04:39:31.807508 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ce8ce318-27e6-434d-96c8-591d250262fd-metrics-certs podName:ce8ce318-27e6-434d-96c8-591d250262fd nodeName:}" failed. No retries permitted until 2025-11-29 04:39:39.807477971 +0000 UTC m=+55.450408401 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ce8ce318-27e6-434d-96c8-591d250262fd-metrics-certs") pod "network-metrics-daemon-qx9wn" (UID: "ce8ce318-27e6-434d-96c8-591d250262fd") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.830591 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.830669 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.830695 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.830731 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.830753 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:31Z","lastTransitionTime":"2025-11-29T04:39:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.934282 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.934366 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.934392 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.934425 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:31 crc kubenswrapper[4799]: I1129 04:39:31.934453 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:31Z","lastTransitionTime":"2025-11-29T04:39:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.038138 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.038254 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.038279 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.038316 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.038344 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:32Z","lastTransitionTime":"2025-11-29T04:39:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.146085 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.146906 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.147021 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.147067 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.147095 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:32Z","lastTransitionTime":"2025-11-29T04:39:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.250685 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.250752 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.250773 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.250839 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.250863 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:32Z","lastTransitionTime":"2025-11-29T04:39:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.354973 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.355051 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.355070 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.355098 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.355120 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:32Z","lastTransitionTime":"2025-11-29T04:39:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.457881 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.457947 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.457964 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.457990 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.458007 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:32Z","lastTransitionTime":"2025-11-29T04:39:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.561598 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.561686 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.561706 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.561736 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.561755 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:32Z","lastTransitionTime":"2025-11-29T04:39:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.664160 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.664207 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.664217 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.664235 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.664264 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:32Z","lastTransitionTime":"2025-11-29T04:39:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.767691 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.767735 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.767746 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.767762 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.767774 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:32Z","lastTransitionTime":"2025-11-29T04:39:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.777754 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.785916 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.798139 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://deb9f370792598608cf2e0c0fe9975eaf8b95dc94b12eef8c55112df4aa9c998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:32Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.811197 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:32Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.825023 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee17b4a9a8999e5f76ceaf69dbd13acf80ed0e95e72f3e1428ce38c5d5982788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://382256e0fa1a4c26610227b9ae96288ccbfbbd97aa3cbd5fd4f18cb4228a234b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:32Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.839432 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt2dl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6c9236-5f7d-46c5-b080-4667620d874e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b6c827d0c57a87808fef301661fce28c50359e85c9e47947c50945f4f869de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt2dl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:32Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.860825 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s5d7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0af3d308-36b9-4622-8544-635f7b24d687\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23c8830e316099ef7b0e0044145a128658456871418c8c53bcc71210be290df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0088b60069e830e9c36075b09d1ff135da79e94fe6cd271d0e93415af50557e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0088b60069e830e9c36075b09d1ff135da79e94fe6cd271d0e93415af50557e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9b0b72fbe8b552d1266c310f6a6a67822c4c70dec1f812d31cc00748d50565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9b0b72fbe8b552d1266c310f6a6a67822c4c70dec1f812d31cc00748d50565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa3b85ae7c47440f0f61436e1cb1f3f2096e0e4011434876f9d17e913ca24df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa3b85ae7c47440f0f61436e1cb1f3f2096e0e4011434876f9d17e913ca24df7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s5d7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:32Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.869985 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.870053 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.870067 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.870086 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.870098 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:32Z","lastTransitionTime":"2025-11-29T04:39:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.879649 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d87rr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a92ae69-5675-49d9-a567-e2476fe02ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://207d152294f9996ed7294d858842da5eef891a6d8a027473dc3d29a7101aadab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffwm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d87rr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:32Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.915151 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://396c1c8cbd638c7dc4d8eebded6a17449d9abdc049161578581f1d8c3e31d9a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:32Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.928608 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v79sj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8cf268-e9fd-460a-b579-6411942a00a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c4e099a3149affd1f9f9c5cbaa0f2758d6ab92117c7352db8ede170153c6bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f943aedb7ef6351aa6b2a46b3e84b6db2fae1145bd40f3b9aee6d897aa0d5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v79sj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:32Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.944019 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d40160e-075f-44b7-9d9d-3291e7774f27\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7079548fe9529583a1c129f57838c24cf06d1cb55a97c05c63a05d0cc2e7ec9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7762237003a5a256ff20a16a4869462e2e0dd1426aa391f7ec74ed9b1f572d4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1129 04:39:02.908489 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1129 04:39:02.908637 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 04:39:02.909677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703886220/tls.crt::/tmp/serving-cert-1703886220/tls.key\\\\\\\"\\\\nI1129 04:39:03.077876 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 04:39:03.080368 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 04:39:03.080390 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 04:39:03.080424 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 04:39:03.080432 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 04:39:03.086120 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 04:39:03.086240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086412 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 04:39:03.086502 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 04:39:03.086563 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 04:39:03.086622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1129 04:39:03.086141 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1129 04:39:03.086836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:38:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:32Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.956013 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:32Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.972020 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.972057 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.972067 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.972081 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.972090 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:32Z","lastTransitionTime":"2025-11-29T04:39:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.976261 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd007c61-4689-4179-95ad-44a9724e8e7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28499112c33332c23be39a11cdf7fc4e1946d04524a3f7c534087179a6620121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dac1c084955d874741d36709fa848c2c475a2d7108d4378b677bb13a88184105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72802532a47b381ead2507e1ae31c49852f9b1d332f3c08fb1e3a972a128eead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c81d94f098efec48c1ff8e172e476bf806ccec5777974eea9c0f3850e5fd40e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3d7f1ba1a408c54f0470fca40e307b70ba208dc7eeece599b684820e3f56c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d91c4f461a1912117cb8ed3a8087c5efaaf0a67455b721bfbeaad1c7dda8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1822ed94255ec78b74e15140fe5391fd667dd39afbda5fab6cdb350b924c833f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1822ed94255ec78b74e15140fe5391fd667dd39afbda5fab6cdb350b924c833f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T04:39:19Z\\\",\\\"message\\\":\\\"890 6259 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1129 04:39:19.793922 6259 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1129 04:39:19.793930 6259 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1129 04:39:19.793939 6259 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1129 04:39:19.793959 6259 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1129 04:39:19.793967 6259 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1129 04:39:19.793998 6259 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1129 04:39:19.794016 6259 factory.go:656] Stopping watch factory\\\\nI1129 04:39:19.794027 6259 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1129 04:39:19.794034 6259 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1129 04:39:19.794040 6259 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1129 04:39:19.794048 6259 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1129 04:39:19.794053 6259 handler.go:208] Removed *v1.Node event handler 7\\\\nI1129 04:39:19.794060 6259 handler.go:208] Removed *v1.Node event handler 2\\\\nI1129 04:39:19.794109 6259 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-fht8l_openshift-ovn-kubernetes(bd007c61-4689-4179-95ad-44a9724e8e7e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d135e42b8886f1f118abf8f4ce4cf2875ac4dc6c148bfc0677cef6b3651562c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fht8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:32Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.988190 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:32Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:32 crc kubenswrapper[4799]: I1129 04:39:32.998221 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7lb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55156232-4c80-45a7-aad5-a5fa2793e4ed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6e5473c80737a5b7b00e5874062e0485ed7a4d74632202090a9b12f331953b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d2dtw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7lb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:32Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.009107 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0847696a7b7926047dc62d75739b3b6a7a04cd58efd1c1c2c992d5ae8d20ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b2ec8f0d384066ba90eb42c1a0a925ade3722b2515b0592d6f2a8f5e533a0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-26t88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:33Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.019407 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-qx9wn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce8ce318-27e6-434d-96c8-591d250262fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c99mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c99mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-qx9wn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:33Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.034365 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22ed28a-3748-4eb9-b3a7-c02aa38f7e94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33c0540ccb097b15f65369ff537aef4ae621cc828ba314eb61f0a0037608cadf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c241d29297a73dce12bd0bc2863c56475eb35b8fe8c93991ba0ef8695e38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f42a55d797c6c94af3a5eac1bf08aa38a82359b7ddd0ba32dcc797218ceee6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebed9ec415bb433935da36df8014f9087236cea9788b2048cbff8714f8ad0df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:33Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.074068 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.074103 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.074112 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.074143 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.074153 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:33Z","lastTransitionTime":"2025-11-29T04:39:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.176430 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.176463 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.176472 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.176485 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.176493 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:33Z","lastTransitionTime":"2025-11-29T04:39:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.278955 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.278985 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.278993 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.279006 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.279014 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:33Z","lastTransitionTime":"2025-11-29T04:39:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.382114 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.382151 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.382161 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.382177 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.382187 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:33Z","lastTransitionTime":"2025-11-29T04:39:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.485714 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.485775 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.485834 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.485881 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.485902 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:33Z","lastTransitionTime":"2025-11-29T04:39:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.589302 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.589377 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.589403 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.589429 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.589447 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:33Z","lastTransitionTime":"2025-11-29T04:39:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.657778 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.657842 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qx9wn" Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.657878 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.657839 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 04:39:33 crc kubenswrapper[4799]: E1129 04:39:33.657963 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 04:39:33 crc kubenswrapper[4799]: E1129 04:39:33.658107 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 04:39:33 crc kubenswrapper[4799]: E1129 04:39:33.658239 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qx9wn" podUID="ce8ce318-27e6-434d-96c8-591d250262fd" Nov 29 04:39:33 crc kubenswrapper[4799]: E1129 04:39:33.658338 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.692429 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.692468 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.692478 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.692496 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.692507 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:33Z","lastTransitionTime":"2025-11-29T04:39:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.795026 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.795103 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.795136 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.795164 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.795186 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:33Z","lastTransitionTime":"2025-11-29T04:39:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.897621 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.897699 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.897723 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.897755 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:33 crc kubenswrapper[4799]: I1129 04:39:33.897780 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:33Z","lastTransitionTime":"2025-11-29T04:39:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.001372 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.001427 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.001443 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.001463 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.001478 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:34Z","lastTransitionTime":"2025-11-29T04:39:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.103887 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.103938 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.103954 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.103972 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.103985 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:34Z","lastTransitionTime":"2025-11-29T04:39:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.206585 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.206659 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.206678 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.206706 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.206727 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:34Z","lastTransitionTime":"2025-11-29T04:39:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.309923 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.309971 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.309984 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.310002 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.310015 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:34Z","lastTransitionTime":"2025-11-29T04:39:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.412920 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.412958 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.412970 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.412986 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.412998 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:34Z","lastTransitionTime":"2025-11-29T04:39:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.515103 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.515140 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.515149 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.515163 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.515172 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:34Z","lastTransitionTime":"2025-11-29T04:39:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.618153 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.618211 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.618229 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.618254 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.618272 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:34Z","lastTransitionTime":"2025-11-29T04:39:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.659567 4799 scope.go:117] "RemoveContainer" containerID="1822ed94255ec78b74e15140fe5391fd667dd39afbda5fab6cdb350b924c833f" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.684724 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:34Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.706071 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee17b4a9a8999e5f76ceaf69dbd13acf80ed0e95e72f3e1428ce38c5d5982788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://382256e0fa1a4c26610227b9ae96288ccbfbbd97aa3cbd5fd4f18cb4228a234b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:34Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.720766 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.720866 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.720879 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.720899 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.720911 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:34Z","lastTransitionTime":"2025-11-29T04:39:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.723630 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt2dl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6c9236-5f7d-46c5-b080-4667620d874e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b6c827d0c57a87808fef301661fce28c50359e85c9e47947c50945f4f869de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt2dl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:34Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.738648 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s5d7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0af3d308-36b9-4622-8544-635f7b24d687\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23c8830e316099ef7b0e0044145a128658456871418c8c53bcc71210be290df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0088b60069e830e9c36075b09d1ff135da79e94fe6cd271d0e93415af50557e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0088b60069e830e9c36075b09d1ff135da79e94fe6cd271d0e93415af50557e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9b0b72fbe8b552d1266c310f6a6a67822c4c70dec1f812d31cc00748d50565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9b0b72fbe8b552d1266c310f6a6a67822c4c70dec1f812d31cc00748d50565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa3b85ae7c47440f0f61436e1cb1f3f2096e0e4011434876f9d17e913ca24df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa3b85ae7c47440f0f61436e1cb1f3f2096e0e4011434876f9d17e913ca24df7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s5d7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:34Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.754974 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d87rr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a92ae69-5675-49d9-a567-e2476fe02ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://207d152294f9996ed7294d858842da5eef891a6d8a027473dc3d29a7101aadab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffwm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d87rr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:34Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.778317 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://deb9f370792598608cf2e0c0fe9975eaf8b95dc94b12eef8c55112df4aa9c998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:34Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.800463 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v79sj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8cf268-e9fd-460a-b579-6411942a00a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c4e099a3149affd1f9f9c5cbaa0f2758d6ab92117c7352db8ede170153c6bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f943aedb7ef6351aa6b2a46b3e84b6db2fae1145bd40f3b9aee6d897aa0d5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v79sj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:34Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.818870 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d40160e-075f-44b7-9d9d-3291e7774f27\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7079548fe9529583a1c129f57838c24cf06d1cb55a97c05c63a05d0cc2e7ec9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7762237003a5a256ff20a16a4869462e2e0dd1426aa391f7ec74ed9b1f572d4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1129 04:39:02.908489 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1129 04:39:02.908637 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 04:39:02.909677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703886220/tls.crt::/tmp/serving-cert-1703886220/tls.key\\\\\\\"\\\\nI1129 04:39:03.077876 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 04:39:03.080368 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 04:39:03.080390 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 04:39:03.080424 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 04:39:03.080432 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 04:39:03.086120 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 04:39:03.086240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086412 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 04:39:03.086502 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 04:39:03.086563 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 04:39:03.086622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1129 04:39:03.086141 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1129 04:39:03.086836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:38:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:34Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.824057 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.824097 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.824109 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.824131 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.824144 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:34Z","lastTransitionTime":"2025-11-29T04:39:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.834426 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://396c1c8cbd638c7dc4d8eebded6a17449d9abdc049161578581f1d8c3e31d9a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:34Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.854255 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd007c61-4689-4179-95ad-44a9724e8e7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28499112c33332c23be39a11cdf7fc4e1946d04524a3f7c534087179a6620121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dac1c084955d874741d36709fa848c2c475a2d7108d4378b677bb13a88184105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72802532a47b381ead2507e1ae31c49852f9b1d332f3c08fb1e3a972a128eead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c81d94f098efec48c1ff8e172e476bf806ccec5777974eea9c0f3850e5fd40e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3d7f1ba1a408c54f0470fca40e307b70ba208dc7eeece599b684820e3f56c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d91c4f461a1912117cb8ed3a8087c5efaaf0a67455b721bfbeaad1c7dda8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1822ed94255ec78b74e15140fe5391fd667dd39afbda5fab6cdb350b924c833f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1822ed94255ec78b74e15140fe5391fd667dd39afbda5fab6cdb350b924c833f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T04:39:19Z\\\",\\\"message\\\":\\\"890 6259 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1129 04:39:19.793922 6259 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1129 04:39:19.793930 6259 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1129 04:39:19.793939 6259 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1129 04:39:19.793959 6259 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1129 04:39:19.793967 6259 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1129 04:39:19.793998 6259 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1129 04:39:19.794016 6259 factory.go:656] Stopping watch factory\\\\nI1129 04:39:19.794027 6259 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1129 04:39:19.794034 6259 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1129 04:39:19.794040 6259 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1129 04:39:19.794048 6259 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1129 04:39:19.794053 6259 handler.go:208] Removed *v1.Node event handler 7\\\\nI1129 04:39:19.794060 6259 handler.go:208] Removed *v1.Node event handler 2\\\\nI1129 04:39:19.794109 6259 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-fht8l_openshift-ovn-kubernetes(bd007c61-4689-4179-95ad-44a9724e8e7e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d135e42b8886f1f118abf8f4ce4cf2875ac4dc6c148bfc0677cef6b3651562c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fht8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:34Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.870849 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82a40839-9eab-4511-8494-ee3d2c6ba0b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://186c216d29cdc56583cd5f39acc7cfd0558f60c668b1f5cbf039b566c1747e12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02c75c8691748a626cbef2de964cb1f4579e702276e8b02d805db0876a8fee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://624c4a09beeecd568d6045a720dfa28c15821d931021cedd8776da4bb36b557f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5b8064433d389b05e3e4744ef4c1f72512db615eae94e2a847bbd4cb2518f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d5b8064433d389b05e3e4744ef4c1f72512db615eae94e2a847bbd4cb2518f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:38:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:34Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.890097 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:34Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.907228 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:34Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.921082 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7lb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55156232-4c80-45a7-aad5-a5fa2793e4ed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6e5473c80737a5b7b00e5874062e0485ed7a4d74632202090a9b12f331953b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d2dtw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7lb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:34Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.926317 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.926353 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.926366 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.926383 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.926394 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:34Z","lastTransitionTime":"2025-11-29T04:39:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.935044 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0847696a7b7926047dc62d75739b3b6a7a04cd58efd1c1c2c992d5ae8d20ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b2ec8f0d384066ba90eb42c1a0a925ade3722b2515b0592d6f2a8f5e533a0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-26t88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:34Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.949838 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-qx9wn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce8ce318-27e6-434d-96c8-591d250262fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c99mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c99mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-qx9wn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:34Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.967153 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22ed28a-3748-4eb9-b3a7-c02aa38f7e94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33c0540ccb097b15f65369ff537aef4ae621cc828ba314eb61f0a0037608cadf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c241d29297a73dce12bd0bc2863c56475eb35b8fe8c93991ba0ef8695e38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f42a55d797c6c94af3a5eac1bf08aa38a82359b7ddd0ba32dcc797218ceee6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebed9ec415bb433935da36df8014f9087236cea9788b2048cbff8714f8ad0df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:34Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.969781 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fht8l_bd007c61-4689-4179-95ad-44a9724e8e7e/ovnkube-controller/1.log" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.973419 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" event={"ID":"bd007c61-4689-4179-95ad-44a9724e8e7e","Type":"ContainerStarted","Data":"d5ce3c68997bc5859df929b844d823adf00352454bb278d57e90edcc44817cbe"} Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.973574 4799 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 29 04:39:34 crc kubenswrapper[4799]: I1129 04:39:34.990107 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82a40839-9eab-4511-8494-ee3d2c6ba0b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://186c216d29cdc56583cd5f39acc7cfd0558f60c668b1f5cbf039b566c1747e12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02c75c8691748a626cbef2de964cb1f4579e702276e8b02d805db0876a8fee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://624c4a09beeecd568d6045a720dfa28c15821d931021cedd8776da4bb36b557f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5b8064433d389b05e3e4744ef4c1f72512db615eae94e2a847bbd4cb2518f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d5b8064433d389b05e3e4744ef4c1f72512db615eae94e2a847bbd4cb2518f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:38:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:34Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.005058 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:35Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.029945 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.030021 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.030041 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.030071 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.030093 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:35Z","lastTransitionTime":"2025-11-29T04:39:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.037428 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd007c61-4689-4179-95ad-44a9724e8e7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28499112c33332c23be39a11cdf7fc4e1946d04524a3f7c534087179a6620121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dac1c084955d874741d36709fa848c2c475a2d7108d4378b677bb13a88184105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72802532a47b381ead2507e1ae31c49852f9b1d332f3c08fb1e3a972a128eead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c81d94f098efec48c1ff8e172e476bf806ccec5777974eea9c0f3850e5fd40e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3d7f1ba1a408c54f0470fca40e307b70ba208dc7eeece599b684820e3f56c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d91c4f461a1912117cb8ed3a8087c5efaaf0a67455b721bfbeaad1c7dda8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5ce3c68997bc5859df929b844d823adf00352454bb278d57e90edcc44817cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1822ed94255ec78b74e15140fe5391fd667dd39afbda5fab6cdb350b924c833f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T04:39:19Z\\\",\\\"message\\\":\\\"890 6259 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1129 04:39:19.793922 6259 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1129 04:39:19.793930 6259 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1129 04:39:19.793939 6259 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1129 04:39:19.793959 6259 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1129 04:39:19.793967 6259 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1129 04:39:19.793998 6259 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1129 04:39:19.794016 6259 factory.go:656] Stopping watch factory\\\\nI1129 04:39:19.794027 6259 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1129 04:39:19.794034 6259 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1129 04:39:19.794040 6259 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1129 04:39:19.794048 6259 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1129 04:39:19.794053 6259 handler.go:208] Removed *v1.Node event handler 7\\\\nI1129 04:39:19.794060 6259 handler.go:208] Removed *v1.Node event handler 2\\\\nI1129 04:39:19.794109 6259 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d135e42b8886f1f118abf8f4ce4cf2875ac4dc6c148bfc0677cef6b3651562c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fht8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:35Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.057652 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22ed28a-3748-4eb9-b3a7-c02aa38f7e94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33c0540ccb097b15f65369ff537aef4ae621cc828ba314eb61f0a0037608cadf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c241d29297a73dce12bd0bc2863c56475eb35b8fe8c93991ba0ef8695e38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f42a55d797c6c94af3a5eac1bf08aa38a82359b7ddd0ba32dcc797218ceee6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebed9ec415bb433935da36df8014f9087236cea9788b2048cbff8714f8ad0df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:35Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.078685 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:35Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.099939 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7lb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55156232-4c80-45a7-aad5-a5fa2793e4ed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6e5473c80737a5b7b00e5874062e0485ed7a4d74632202090a9b12f331953b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d2dtw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7lb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:35Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.122187 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0847696a7b7926047dc62d75739b3b6a7a04cd58efd1c1c2c992d5ae8d20ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b2ec8f0d384066ba90eb42c1a0a925ade3722b2515b0592d6f2a8f5e533a0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-26t88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:35Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.132524 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.132558 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.132600 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.132617 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.132629 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:35Z","lastTransitionTime":"2025-11-29T04:39:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.146203 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-qx9wn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce8ce318-27e6-434d-96c8-591d250262fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c99mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c99mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-qx9wn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:35Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.161817 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d87rr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a92ae69-5675-49d9-a567-e2476fe02ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://207d152294f9996ed7294d858842da5eef891a6d8a027473dc3d29a7101aadab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffwm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d87rr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:35Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.183868 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://deb9f370792598608cf2e0c0fe9975eaf8b95dc94b12eef8c55112df4aa9c998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:35Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.230573 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:35Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.234572 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.234603 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.234631 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.234645 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.234656 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:35Z","lastTransitionTime":"2025-11-29T04:39:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.242667 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee17b4a9a8999e5f76ceaf69dbd13acf80ed0e95e72f3e1428ce38c5d5982788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://382256e0fa1a4c26610227b9ae96288ccbfbbd97aa3cbd5fd4f18cb4228a234b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:35Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.256839 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt2dl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6c9236-5f7d-46c5-b080-4667620d874e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b6c827d0c57a87808fef301661fce28c50359e85c9e47947c50945f4f869de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt2dl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:35Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.274759 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s5d7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0af3d308-36b9-4622-8544-635f7b24d687\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23c8830e316099ef7b0e0044145a128658456871418c8c53bcc71210be290df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0088b60069e830e9c36075b09d1ff135da79e94fe6cd271d0e93415af50557e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0088b60069e830e9c36075b09d1ff135da79e94fe6cd271d0e93415af50557e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9b0b72fbe8b552d1266c310f6a6a67822c4c70dec1f812d31cc00748d50565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9b0b72fbe8b552d1266c310f6a6a67822c4c70dec1f812d31cc00748d50565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa3b85ae7c47440f0f61436e1cb1f3f2096e0e4011434876f9d17e913ca24df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa3b85ae7c47440f0f61436e1cb1f3f2096e0e4011434876f9d17e913ca24df7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s5d7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:35Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.291835 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d40160e-075f-44b7-9d9d-3291e7774f27\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7079548fe9529583a1c129f57838c24cf06d1cb55a97c05c63a05d0cc2e7ec9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7762237003a5a256ff20a16a4869462e2e0dd1426aa391f7ec74ed9b1f572d4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1129 04:39:02.908489 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1129 04:39:02.908637 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 04:39:02.909677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703886220/tls.crt::/tmp/serving-cert-1703886220/tls.key\\\\\\\"\\\\nI1129 04:39:03.077876 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 04:39:03.080368 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 04:39:03.080390 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 04:39:03.080424 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 04:39:03.080432 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 04:39:03.086120 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 04:39:03.086240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086412 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 04:39:03.086502 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 04:39:03.086563 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 04:39:03.086622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1129 04:39:03.086141 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1129 04:39:03.086836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:38:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:35Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.305168 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://396c1c8cbd638c7dc4d8eebded6a17449d9abdc049161578581f1d8c3e31d9a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:35Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.318184 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v79sj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8cf268-e9fd-460a-b579-6411942a00a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c4e099a3149affd1f9f9c5cbaa0f2758d6ab92117c7352db8ede170153c6bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f943aedb7ef6351aa6b2a46b3e84b6db2fae1145bd40f3b9aee6d897aa0d5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v79sj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:35Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.336823 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.336877 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.336889 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.336907 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.336919 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:35Z","lastTransitionTime":"2025-11-29T04:39:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.440382 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.440418 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.440427 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.440440 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.440449 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:35Z","lastTransitionTime":"2025-11-29T04:39:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.543555 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.543593 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.543602 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.543618 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.543629 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:35Z","lastTransitionTime":"2025-11-29T04:39:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.651970 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 04:39:35 crc kubenswrapper[4799]: E1129 04:39:35.652499 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 04:40:07.652364875 +0000 UTC m=+83.295295315 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.653706 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.653897 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.654014 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.654041 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.653917 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 04:39:35 crc kubenswrapper[4799]: E1129 04:39:35.653932 4799 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 29 04:39:35 crc kubenswrapper[4799]: E1129 04:39:35.654119 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 29 04:39:35 crc kubenswrapper[4799]: E1129 04:39:35.654154 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 29 04:39:35 crc kubenswrapper[4799]: E1129 04:39:35.654178 4799 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.654076 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.654276 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:35Z","lastTransitionTime":"2025-11-29T04:39:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.654850 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:39:35 crc kubenswrapper[4799]: E1129 04:39:35.654941 4799 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 29 04:39:35 crc kubenswrapper[4799]: E1129 04:39:35.654988 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-29 04:40:07.654978909 +0000 UTC m=+83.297909309 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 29 04:39:35 crc kubenswrapper[4799]: E1129 04:39:35.655027 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-29 04:40:07.654995869 +0000 UTC m=+83.297926269 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 29 04:39:35 crc kubenswrapper[4799]: E1129 04:39:35.655061 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-29 04:40:07.655052431 +0000 UTC m=+83.297982831 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.655126 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 04:39:35 crc kubenswrapper[4799]: E1129 04:39:35.655250 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 29 04:39:35 crc kubenswrapper[4799]: E1129 04:39:35.655289 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 29 04:39:35 crc kubenswrapper[4799]: E1129 04:39:35.655306 4799 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 04:39:35 crc kubenswrapper[4799]: E1129 04:39:35.655382 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-29 04:40:07.655352818 +0000 UTC m=+83.298283258 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.657892 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 04:39:35 crc kubenswrapper[4799]: E1129 04:39:35.658055 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.658120 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qx9wn" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.658161 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 04:39:35 crc kubenswrapper[4799]: E1129 04:39:35.658229 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qx9wn" podUID="ce8ce318-27e6-434d-96c8-591d250262fd" Nov 29 04:39:35 crc kubenswrapper[4799]: E1129 04:39:35.658379 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.658570 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:39:35 crc kubenswrapper[4799]: E1129 04:39:35.658736 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.757877 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.757912 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.757921 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.757936 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.757946 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:35Z","lastTransitionTime":"2025-11-29T04:39:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.860655 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.860739 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.860766 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.860829 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.860860 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:35Z","lastTransitionTime":"2025-11-29T04:39:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.965461 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.965545 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.965568 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.965599 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.965621 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:35Z","lastTransitionTime":"2025-11-29T04:39:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.980680 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fht8l_bd007c61-4689-4179-95ad-44a9724e8e7e/ovnkube-controller/2.log" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.981764 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fht8l_bd007c61-4689-4179-95ad-44a9724e8e7e/ovnkube-controller/1.log" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.985985 4799 generic.go:334] "Generic (PLEG): container finished" podID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerID="d5ce3c68997bc5859df929b844d823adf00352454bb278d57e90edcc44817cbe" exitCode=1 Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.986048 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" event={"ID":"bd007c61-4689-4179-95ad-44a9724e8e7e","Type":"ContainerDied","Data":"d5ce3c68997bc5859df929b844d823adf00352454bb278d57e90edcc44817cbe"} Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.986109 4799 scope.go:117] "RemoveContainer" containerID="1822ed94255ec78b74e15140fe5391fd667dd39afbda5fab6cdb350b924c833f" Nov 29 04:39:35 crc kubenswrapper[4799]: I1129 04:39:35.987701 4799 scope.go:117] "RemoveContainer" containerID="d5ce3c68997bc5859df929b844d823adf00352454bb278d57e90edcc44817cbe" Nov 29 04:39:35 crc kubenswrapper[4799]: E1129 04:39:35.988125 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fht8l_openshift-ovn-kubernetes(bd007c61-4689-4179-95ad-44a9724e8e7e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" podUID="bd007c61-4689-4179-95ad-44a9724e8e7e" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.010474 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d40160e-075f-44b7-9d9d-3291e7774f27\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7079548fe9529583a1c129f57838c24cf06d1cb55a97c05c63a05d0cc2e7ec9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7762237003a5a256ff20a16a4869462e2e0dd1426aa391f7ec74ed9b1f572d4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1129 04:39:02.908489 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1129 04:39:02.908637 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 04:39:02.909677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703886220/tls.crt::/tmp/serving-cert-1703886220/tls.key\\\\\\\"\\\\nI1129 04:39:03.077876 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 04:39:03.080368 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 04:39:03.080390 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 04:39:03.080424 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 04:39:03.080432 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 04:39:03.086120 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 04:39:03.086240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086412 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 04:39:03.086502 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 04:39:03.086563 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 04:39:03.086622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1129 04:39:03.086141 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1129 04:39:03.086836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:38:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:36Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.026476 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://396c1c8cbd638c7dc4d8eebded6a17449d9abdc049161578581f1d8c3e31d9a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:36Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.039743 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v79sj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8cf268-e9fd-460a-b579-6411942a00a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c4e099a3149affd1f9f9c5cbaa0f2758d6ab92117c7352db8ede170153c6bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f943aedb7ef6351aa6b2a46b3e84b6db2fae1145bd40f3b9aee6d897aa0d5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v79sj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:36Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.054055 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82a40839-9eab-4511-8494-ee3d2c6ba0b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://186c216d29cdc56583cd5f39acc7cfd0558f60c668b1f5cbf039b566c1747e12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02c75c8691748a626cbef2de964cb1f4579e702276e8b02d805db0876a8fee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://624c4a09beeecd568d6045a720dfa28c15821d931021cedd8776da4bb36b557f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5b8064433d389b05e3e4744ef4c1f72512db615eae94e2a847bbd4cb2518f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d5b8064433d389b05e3e4744ef4c1f72512db615eae94e2a847bbd4cb2518f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:38:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:36Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.069587 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.069634 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.069644 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.069666 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.069678 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:36Z","lastTransitionTime":"2025-11-29T04:39:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.076315 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:36Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.097137 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd007c61-4689-4179-95ad-44a9724e8e7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28499112c33332c23be39a11cdf7fc4e1946d04524a3f7c534087179a6620121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dac1c084955d874741d36709fa848c2c475a2d7108d4378b677bb13a88184105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72802532a47b381ead2507e1ae31c49852f9b1d332f3c08fb1e3a972a128eead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c81d94f098efec48c1ff8e172e476bf806ccec5777974eea9c0f3850e5fd40e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3d7f1ba1a408c54f0470fca40e307b70ba208dc7eeece599b684820e3f56c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d91c4f461a1912117cb8ed3a8087c5efaaf0a67455b721bfbeaad1c7dda8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5ce3c68997bc5859df929b844d823adf00352454bb278d57e90edcc44817cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1822ed94255ec78b74e15140fe5391fd667dd39afbda5fab6cdb350b924c833f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T04:39:19Z\\\",\\\"message\\\":\\\"890 6259 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1129 04:39:19.793922 6259 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1129 04:39:19.793930 6259 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1129 04:39:19.793939 6259 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1129 04:39:19.793959 6259 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1129 04:39:19.793967 6259 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1129 04:39:19.793998 6259 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1129 04:39:19.794016 6259 factory.go:656] Stopping watch factory\\\\nI1129 04:39:19.794027 6259 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1129 04:39:19.794034 6259 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1129 04:39:19.794040 6259 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1129 04:39:19.794048 6259 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1129 04:39:19.794053 6259 handler.go:208] Removed *v1.Node event handler 7\\\\nI1129 04:39:19.794060 6259 handler.go:208] Removed *v1.Node event handler 2\\\\nI1129 04:39:19.794109 6259 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5ce3c68997bc5859df929b844d823adf00352454bb278d57e90edcc44817cbe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T04:39:35Z\\\",\\\"message\\\":\\\"l\\\\u003e UUID: UUIDName:} {Op:mutate Table:NB_Global Row:map[] Rows:[] Columns:[] Mutations:[{Column:nb_cfg Mutator:+= Value:1}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {6011affd-30a6-4be6-872d-e4cf1ca780cf}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1129 04:39:35.599606 6479 services_controller.go:434] Service openshift-multus/multus-admission-controller retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{multus-admission-controller openshift-multus c9eea3b1-f918-4c62-9731-c809988317c1 4579 0 2025-02-23 05:21:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:multus-admission-controller] map[service.alpha.openshift.io/serving-cert-secret-name:multus-admission-controller-secret service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{operator.openshift.io/v1 Network cluster 8d01ddba-7e05-4639-926a-4485de3b6327 0xc007916917 0xc007916918}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:webhook,Protocol:TCP,Port:443,TargetPort:{0 6443 },NodePort:0,AppProtocol:nil,},ServicePort{Name:metrics,Protocol:TCP,Port:8443,TargetPort:{1 0 https},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: multus-admission-controller,},ClusterIP:10.217.5.119,Type:ClusterIP,ExternalIPs:[],Sessi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d135e42b8886f1f118abf8f4ce4cf2875ac4dc6c148bfc0677cef6b3651562c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fht8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:36Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.116041 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22ed28a-3748-4eb9-b3a7-c02aa38f7e94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33c0540ccb097b15f65369ff537aef4ae621cc828ba314eb61f0a0037608cadf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c241d29297a73dce12bd0bc2863c56475eb35b8fe8c93991ba0ef8695e38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f42a55d797c6c94af3a5eac1bf08aa38a82359b7ddd0ba32dcc797218ceee6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebed9ec415bb433935da36df8014f9087236cea9788b2048cbff8714f8ad0df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:36Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.137832 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:36Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.155572 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7lb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55156232-4c80-45a7-aad5-a5fa2793e4ed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6e5473c80737a5b7b00e5874062e0485ed7a4d74632202090a9b12f331953b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d2dtw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7lb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:36Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.173476 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.173540 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.173558 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.173586 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.173604 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:36Z","lastTransitionTime":"2025-11-29T04:39:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.174722 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0847696a7b7926047dc62d75739b3b6a7a04cd58efd1c1c2c992d5ae8d20ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b2ec8f0d384066ba90eb42c1a0a925ade3722b2515b0592d6f2a8f5e533a0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-26t88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:36Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.194036 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-qx9wn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce8ce318-27e6-434d-96c8-591d250262fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c99mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c99mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-qx9wn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:36Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.208717 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d87rr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a92ae69-5675-49d9-a567-e2476fe02ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://207d152294f9996ed7294d858842da5eef891a6d8a027473dc3d29a7101aadab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffwm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d87rr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:36Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.225098 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://deb9f370792598608cf2e0c0fe9975eaf8b95dc94b12eef8c55112df4aa9c998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:36Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.242103 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:36Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.260412 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee17b4a9a8999e5f76ceaf69dbd13acf80ed0e95e72f3e1428ce38c5d5982788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://382256e0fa1a4c26610227b9ae96288ccbfbbd97aa3cbd5fd4f18cb4228a234b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:36Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.275835 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt2dl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6c9236-5f7d-46c5-b080-4667620d874e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b6c827d0c57a87808fef301661fce28c50359e85c9e47947c50945f4f869de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt2dl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:36Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.276157 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.276201 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.276213 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.276231 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.276243 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:36Z","lastTransitionTime":"2025-11-29T04:39:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.295503 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s5d7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0af3d308-36b9-4622-8544-635f7b24d687\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23c8830e316099ef7b0e0044145a128658456871418c8c53bcc71210be290df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0088b60069e830e9c36075b09d1ff135da79e94fe6cd271d0e93415af50557e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0088b60069e830e9c36075b09d1ff135da79e94fe6cd271d0e93415af50557e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9b0b72fbe8b552d1266c310f6a6a67822c4c70dec1f812d31cc00748d50565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9b0b72fbe8b552d1266c310f6a6a67822c4c70dec1f812d31cc00748d50565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa3b85ae7c47440f0f61436e1cb1f3f2096e0e4011434876f9d17e913ca24df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa3b85ae7c47440f0f61436e1cb1f3f2096e0e4011434876f9d17e913ca24df7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s5d7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:36Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.378725 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.379031 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.379228 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.379324 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.379386 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:36Z","lastTransitionTime":"2025-11-29T04:39:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.481935 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.481970 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.481979 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.481993 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.482001 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:36Z","lastTransitionTime":"2025-11-29T04:39:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.584424 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.584479 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.584488 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.584501 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.584510 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:36Z","lastTransitionTime":"2025-11-29T04:39:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.685926 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.685963 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.685972 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.685986 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.685997 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:36Z","lastTransitionTime":"2025-11-29T04:39:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.789122 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.789167 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.789179 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.789197 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.789242 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:36Z","lastTransitionTime":"2025-11-29T04:39:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.892286 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.892323 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.892331 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.892347 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.892357 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:36Z","lastTransitionTime":"2025-11-29T04:39:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.990034 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fht8l_bd007c61-4689-4179-95ad-44a9724e8e7e/ovnkube-controller/2.log" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.993783 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.993855 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.993898 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.993924 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:36 crc kubenswrapper[4799]: I1129 04:39:36.993935 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:36Z","lastTransitionTime":"2025-11-29T04:39:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:37 crc kubenswrapper[4799]: I1129 04:39:37.096274 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:37 crc kubenswrapper[4799]: I1129 04:39:37.096319 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:37 crc kubenswrapper[4799]: I1129 04:39:37.096355 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:37 crc kubenswrapper[4799]: I1129 04:39:37.096370 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:37 crc kubenswrapper[4799]: I1129 04:39:37.096381 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:37Z","lastTransitionTime":"2025-11-29T04:39:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:37 crc kubenswrapper[4799]: I1129 04:39:37.198695 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:37 crc kubenswrapper[4799]: I1129 04:39:37.198734 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:37 crc kubenswrapper[4799]: I1129 04:39:37.198744 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:37 crc kubenswrapper[4799]: I1129 04:39:37.198759 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:37 crc kubenswrapper[4799]: I1129 04:39:37.198769 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:37Z","lastTransitionTime":"2025-11-29T04:39:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:37 crc kubenswrapper[4799]: I1129 04:39:37.301753 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:37 crc kubenswrapper[4799]: I1129 04:39:37.301872 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:37 crc kubenswrapper[4799]: I1129 04:39:37.301895 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:37 crc kubenswrapper[4799]: I1129 04:39:37.301917 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:37 crc kubenswrapper[4799]: I1129 04:39:37.301931 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:37Z","lastTransitionTime":"2025-11-29T04:39:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:37 crc kubenswrapper[4799]: I1129 04:39:37.403969 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:37 crc kubenswrapper[4799]: I1129 04:39:37.404006 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:37 crc kubenswrapper[4799]: I1129 04:39:37.404016 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:37 crc kubenswrapper[4799]: I1129 04:39:37.404033 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:37 crc kubenswrapper[4799]: I1129 04:39:37.404042 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:37Z","lastTransitionTime":"2025-11-29T04:39:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:37 crc kubenswrapper[4799]: I1129 04:39:37.506779 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:37 crc kubenswrapper[4799]: I1129 04:39:37.506942 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:37 crc kubenswrapper[4799]: I1129 04:39:37.506964 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:37 crc kubenswrapper[4799]: I1129 04:39:37.506997 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:37 crc kubenswrapper[4799]: I1129 04:39:37.507021 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:37Z","lastTransitionTime":"2025-11-29T04:39:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:37 crc kubenswrapper[4799]: I1129 04:39:37.609835 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:37 crc kubenswrapper[4799]: I1129 04:39:37.609872 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:37 crc kubenswrapper[4799]: I1129 04:39:37.609883 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:37 crc kubenswrapper[4799]: I1129 04:39:37.609898 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:37 crc kubenswrapper[4799]: I1129 04:39:37.609916 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:37Z","lastTransitionTime":"2025-11-29T04:39:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:37 crc kubenswrapper[4799]: I1129 04:39:37.658362 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:39:37 crc kubenswrapper[4799]: I1129 04:39:37.658421 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qx9wn" Nov 29 04:39:37 crc kubenswrapper[4799]: I1129 04:39:37.658362 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 04:39:37 crc kubenswrapper[4799]: I1129 04:39:37.658504 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 04:39:37 crc kubenswrapper[4799]: E1129 04:39:37.658660 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qx9wn" podUID="ce8ce318-27e6-434d-96c8-591d250262fd" Nov 29 04:39:37 crc kubenswrapper[4799]: E1129 04:39:37.658769 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 04:39:37 crc kubenswrapper[4799]: E1129 04:39:37.658930 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 04:39:37 crc kubenswrapper[4799]: E1129 04:39:37.659034 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 04:39:37 crc kubenswrapper[4799]: I1129 04:39:37.712998 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:37 crc kubenswrapper[4799]: I1129 04:39:37.713074 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:37 crc kubenswrapper[4799]: I1129 04:39:37.713098 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:37 crc kubenswrapper[4799]: I1129 04:39:37.713129 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:37 crc kubenswrapper[4799]: I1129 04:39:37.713151 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:37Z","lastTransitionTime":"2025-11-29T04:39:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:37 crc kubenswrapper[4799]: I1129 04:39:37.816368 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:37 crc kubenswrapper[4799]: I1129 04:39:37.816476 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:37 crc kubenswrapper[4799]: I1129 04:39:37.816533 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:37 crc kubenswrapper[4799]: I1129 04:39:37.816570 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:37 crc kubenswrapper[4799]: I1129 04:39:37.816599 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:37Z","lastTransitionTime":"2025-11-29T04:39:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:37 crc kubenswrapper[4799]: I1129 04:39:37.919160 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:37 crc kubenswrapper[4799]: I1129 04:39:37.919192 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:37 crc kubenswrapper[4799]: I1129 04:39:37.919201 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:37 crc kubenswrapper[4799]: I1129 04:39:37.919218 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:37 crc kubenswrapper[4799]: I1129 04:39:37.919227 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:37Z","lastTransitionTime":"2025-11-29T04:39:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.021740 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.021786 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.021823 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.021847 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.021870 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:38Z","lastTransitionTime":"2025-11-29T04:39:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.124943 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.125013 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.125031 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.125056 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.125075 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:38Z","lastTransitionTime":"2025-11-29T04:39:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.228417 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.228485 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.228502 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.228530 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.228552 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:38Z","lastTransitionTime":"2025-11-29T04:39:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.331483 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.331535 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.331552 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.331577 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.331594 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:38Z","lastTransitionTime":"2025-11-29T04:39:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.434103 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.434192 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.434215 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.434242 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.434267 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:38Z","lastTransitionTime":"2025-11-29T04:39:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.536187 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.536217 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.536226 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.536242 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.536253 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:38Z","lastTransitionTime":"2025-11-29T04:39:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.553558 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.553614 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.553626 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.553644 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.553655 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:38Z","lastTransitionTime":"2025-11-29T04:39:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:38 crc kubenswrapper[4799]: E1129 04:39:38.572469 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6f4e4672-da32-440c-8c4f-6e7f4942f55e\\\",\\\"systemUUID\\\":\\\"882c4d8e-57b7-4312-8499-40d1f85d56c9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:38Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.578155 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.578188 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.578196 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.578212 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.578224 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:38Z","lastTransitionTime":"2025-11-29T04:39:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:38 crc kubenswrapper[4799]: E1129 04:39:38.600224 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6f4e4672-da32-440c-8c4f-6e7f4942f55e\\\",\\\"systemUUID\\\":\\\"882c4d8e-57b7-4312-8499-40d1f85d56c9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:38Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.605709 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.605759 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.605769 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.605810 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.605822 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:38Z","lastTransitionTime":"2025-11-29T04:39:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:38 crc kubenswrapper[4799]: E1129 04:39:38.623512 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6f4e4672-da32-440c-8c4f-6e7f4942f55e\\\",\\\"systemUUID\\\":\\\"882c4d8e-57b7-4312-8499-40d1f85d56c9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:38Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.627563 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.627710 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.627824 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.627931 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.628008 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:38Z","lastTransitionTime":"2025-11-29T04:39:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:38 crc kubenswrapper[4799]: E1129 04:39:38.642563 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6f4e4672-da32-440c-8c4f-6e7f4942f55e\\\",\\\"systemUUID\\\":\\\"882c4d8e-57b7-4312-8499-40d1f85d56c9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:38Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.648243 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.648331 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.648357 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.648389 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.648409 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:38Z","lastTransitionTime":"2025-11-29T04:39:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:38 crc kubenswrapper[4799]: E1129 04:39:38.669729 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6f4e4672-da32-440c-8c4f-6e7f4942f55e\\\",\\\"systemUUID\\\":\\\"882c4d8e-57b7-4312-8499-40d1f85d56c9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:38Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:38 crc kubenswrapper[4799]: E1129 04:39:38.670012 4799 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.671733 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.671774 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.671824 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.671849 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.671869 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:38Z","lastTransitionTime":"2025-11-29T04:39:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.774441 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.774486 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.774498 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.774515 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.774530 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:38Z","lastTransitionTime":"2025-11-29T04:39:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.877886 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.877944 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.877957 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.877977 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.877990 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:38Z","lastTransitionTime":"2025-11-29T04:39:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.981546 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.981675 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.981696 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.981736 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:38 crc kubenswrapper[4799]: I1129 04:39:38.981759 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:38Z","lastTransitionTime":"2025-11-29T04:39:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:39 crc kubenswrapper[4799]: I1129 04:39:39.084650 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:39 crc kubenswrapper[4799]: I1129 04:39:39.084690 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:39 crc kubenswrapper[4799]: I1129 04:39:39.084699 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:39 crc kubenswrapper[4799]: I1129 04:39:39.084724 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:39 crc kubenswrapper[4799]: I1129 04:39:39.084734 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:39Z","lastTransitionTime":"2025-11-29T04:39:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:39 crc kubenswrapper[4799]: I1129 04:39:39.186895 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:39 crc kubenswrapper[4799]: I1129 04:39:39.186939 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:39 crc kubenswrapper[4799]: I1129 04:39:39.186966 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:39 crc kubenswrapper[4799]: I1129 04:39:39.186981 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:39 crc kubenswrapper[4799]: I1129 04:39:39.186991 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:39Z","lastTransitionTime":"2025-11-29T04:39:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:39 crc kubenswrapper[4799]: I1129 04:39:39.289298 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:39 crc kubenswrapper[4799]: I1129 04:39:39.289341 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:39 crc kubenswrapper[4799]: I1129 04:39:39.289350 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:39 crc kubenswrapper[4799]: I1129 04:39:39.289367 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:39 crc kubenswrapper[4799]: I1129 04:39:39.289380 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:39Z","lastTransitionTime":"2025-11-29T04:39:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:39 crc kubenswrapper[4799]: I1129 04:39:39.392090 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:39 crc kubenswrapper[4799]: I1129 04:39:39.392178 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:39 crc kubenswrapper[4799]: I1129 04:39:39.392205 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:39 crc kubenswrapper[4799]: I1129 04:39:39.392240 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:39 crc kubenswrapper[4799]: I1129 04:39:39.392262 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:39Z","lastTransitionTime":"2025-11-29T04:39:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:39 crc kubenswrapper[4799]: I1129 04:39:39.495242 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:39 crc kubenswrapper[4799]: I1129 04:39:39.495326 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:39 crc kubenswrapper[4799]: I1129 04:39:39.495350 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:39 crc kubenswrapper[4799]: I1129 04:39:39.495381 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:39 crc kubenswrapper[4799]: I1129 04:39:39.495404 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:39Z","lastTransitionTime":"2025-11-29T04:39:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:39 crc kubenswrapper[4799]: I1129 04:39:39.597552 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:39 crc kubenswrapper[4799]: I1129 04:39:39.597599 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:39 crc kubenswrapper[4799]: I1129 04:39:39.597614 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:39 crc kubenswrapper[4799]: I1129 04:39:39.597639 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:39 crc kubenswrapper[4799]: I1129 04:39:39.597658 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:39Z","lastTransitionTime":"2025-11-29T04:39:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:39 crc kubenswrapper[4799]: I1129 04:39:39.658560 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 04:39:39 crc kubenswrapper[4799]: I1129 04:39:39.658584 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 04:39:39 crc kubenswrapper[4799]: I1129 04:39:39.658650 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:39:39 crc kubenswrapper[4799]: I1129 04:39:39.658610 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qx9wn" Nov 29 04:39:39 crc kubenswrapper[4799]: E1129 04:39:39.658812 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 04:39:39 crc kubenswrapper[4799]: E1129 04:39:39.659045 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 04:39:39 crc kubenswrapper[4799]: E1129 04:39:39.659240 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 04:39:39 crc kubenswrapper[4799]: E1129 04:39:39.659355 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qx9wn" podUID="ce8ce318-27e6-434d-96c8-591d250262fd" Nov 29 04:39:39 crc kubenswrapper[4799]: I1129 04:39:39.700807 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:39 crc kubenswrapper[4799]: I1129 04:39:39.700851 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:39 crc kubenswrapper[4799]: I1129 04:39:39.700864 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:39 crc kubenswrapper[4799]: I1129 04:39:39.700884 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:39 crc kubenswrapper[4799]: I1129 04:39:39.700898 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:39Z","lastTransitionTime":"2025-11-29T04:39:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:39 crc kubenswrapper[4799]: I1129 04:39:39.802804 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:39 crc kubenswrapper[4799]: I1129 04:39:39.802873 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:39 crc kubenswrapper[4799]: I1129 04:39:39.802885 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:39 crc kubenswrapper[4799]: I1129 04:39:39.802909 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:39 crc kubenswrapper[4799]: I1129 04:39:39.802923 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:39Z","lastTransitionTime":"2025-11-29T04:39:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:39 crc kubenswrapper[4799]: I1129 04:39:39.906297 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:39 crc kubenswrapper[4799]: I1129 04:39:39.906380 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:39 crc kubenswrapper[4799]: I1129 04:39:39.906400 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:39 crc kubenswrapper[4799]: I1129 04:39:39.906436 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:39 crc kubenswrapper[4799]: I1129 04:39:39.906457 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:39Z","lastTransitionTime":"2025-11-29T04:39:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:39 crc kubenswrapper[4799]: I1129 04:39:39.906980 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ce8ce318-27e6-434d-96c8-591d250262fd-metrics-certs\") pod \"network-metrics-daemon-qx9wn\" (UID: \"ce8ce318-27e6-434d-96c8-591d250262fd\") " pod="openshift-multus/network-metrics-daemon-qx9wn" Nov 29 04:39:39 crc kubenswrapper[4799]: E1129 04:39:39.907111 4799 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 29 04:39:39 crc kubenswrapper[4799]: E1129 04:39:39.907166 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ce8ce318-27e6-434d-96c8-591d250262fd-metrics-certs podName:ce8ce318-27e6-434d-96c8-591d250262fd nodeName:}" failed. No retries permitted until 2025-11-29 04:39:55.907152802 +0000 UTC m=+71.550083202 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ce8ce318-27e6-434d-96c8-591d250262fd-metrics-certs") pod "network-metrics-daemon-qx9wn" (UID: "ce8ce318-27e6-434d-96c8-591d250262fd") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.009438 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.009480 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.009518 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.009533 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.009544 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:40Z","lastTransitionTime":"2025-11-29T04:39:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.111373 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.111466 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.111484 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.111508 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.111525 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:40Z","lastTransitionTime":"2025-11-29T04:39:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.217091 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.217141 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.217153 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.217169 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.217182 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:40Z","lastTransitionTime":"2025-11-29T04:39:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.319455 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.319727 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.319877 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.319972 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.320062 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:40Z","lastTransitionTime":"2025-11-29T04:39:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.408159 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.409641 4799 scope.go:117] "RemoveContainer" containerID="d5ce3c68997bc5859df929b844d823adf00352454bb278d57e90edcc44817cbe" Nov 29 04:39:40 crc kubenswrapper[4799]: E1129 04:39:40.409941 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fht8l_openshift-ovn-kubernetes(bd007c61-4689-4179-95ad-44a9724e8e7e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" podUID="bd007c61-4689-4179-95ad-44a9724e8e7e" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.422122 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.422167 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.422178 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.422196 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.422207 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:40Z","lastTransitionTime":"2025-11-29T04:39:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.426497 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s5d7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0af3d308-36b9-4622-8544-635f7b24d687\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23c8830e316099ef7b0e0044145a128658456871418c8c53bcc71210be290df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0088b60069e830e9c36075b09d1ff135da79e94fe6cd271d0e93415af50557e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0088b60069e830e9c36075b09d1ff135da79e94fe6cd271d0e93415af50557e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9b0b72fbe8b552d1266c310f6a6a67822c4c70dec1f812d31cc00748d50565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9b0b72fbe8b552d1266c310f6a6a67822c4c70dec1f812d31cc00748d50565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa3b85ae7c47440f0f61436e1cb1f3f2096e0e4011434876f9d17e913ca24df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa3b85ae7c47440f0f61436e1cb1f3f2096e0e4011434876f9d17e913ca24df7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s5d7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:40Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.441001 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d87rr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a92ae69-5675-49d9-a567-e2476fe02ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://207d152294f9996ed7294d858842da5eef891a6d8a027473dc3d29a7101aadab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffwm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d87rr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:40Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.455760 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://deb9f370792598608cf2e0c0fe9975eaf8b95dc94b12eef8c55112df4aa9c998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:40Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.472195 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:40Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.487829 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee17b4a9a8999e5f76ceaf69dbd13acf80ed0e95e72f3e1428ce38c5d5982788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://382256e0fa1a4c26610227b9ae96288ccbfbbd97aa3cbd5fd4f18cb4228a234b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:40Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.505119 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt2dl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6c9236-5f7d-46c5-b080-4667620d874e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b6c827d0c57a87808fef301661fce28c50359e85c9e47947c50945f4f869de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt2dl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:40Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.521372 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d40160e-075f-44b7-9d9d-3291e7774f27\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7079548fe9529583a1c129f57838c24cf06d1cb55a97c05c63a05d0cc2e7ec9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7762237003a5a256ff20a16a4869462e2e0dd1426aa391f7ec74ed9b1f572d4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1129 04:39:02.908489 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1129 04:39:02.908637 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 04:39:02.909677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703886220/tls.crt::/tmp/serving-cert-1703886220/tls.key\\\\\\\"\\\\nI1129 04:39:03.077876 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 04:39:03.080368 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 04:39:03.080390 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 04:39:03.080424 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 04:39:03.080432 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 04:39:03.086120 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 04:39:03.086240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086412 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 04:39:03.086502 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 04:39:03.086563 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 04:39:03.086622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1129 04:39:03.086141 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1129 04:39:03.086836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:38:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:40Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.525177 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.525216 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.525228 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.525244 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.525257 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:40Z","lastTransitionTime":"2025-11-29T04:39:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.535236 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://396c1c8cbd638c7dc4d8eebded6a17449d9abdc049161578581f1d8c3e31d9a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:40Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.547139 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v79sj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8cf268-e9fd-460a-b579-6411942a00a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c4e099a3149affd1f9f9c5cbaa0f2758d6ab92117c7352db8ede170153c6bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f943aedb7ef6351aa6b2a46b3e84b6db2fae1145bd40f3b9aee6d897aa0d5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v79sj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:40Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.558656 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82a40839-9eab-4511-8494-ee3d2c6ba0b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://186c216d29cdc56583cd5f39acc7cfd0558f60c668b1f5cbf039b566c1747e12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02c75c8691748a626cbef2de964cb1f4579e702276e8b02d805db0876a8fee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://624c4a09beeecd568d6045a720dfa28c15821d931021cedd8776da4bb36b557f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5b8064433d389b05e3e4744ef4c1f72512db615eae94e2a847bbd4cb2518f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d5b8064433d389b05e3e4744ef4c1f72512db615eae94e2a847bbd4cb2518f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:38:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:40Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.571778 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:40Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.588497 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd007c61-4689-4179-95ad-44a9724e8e7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28499112c33332c23be39a11cdf7fc4e1946d04524a3f7c534087179a6620121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dac1c084955d874741d36709fa848c2c475a2d7108d4378b677bb13a88184105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72802532a47b381ead2507e1ae31c49852f9b1d332f3c08fb1e3a972a128eead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c81d94f098efec48c1ff8e172e476bf806ccec5777974eea9c0f3850e5fd40e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3d7f1ba1a408c54f0470fca40e307b70ba208dc7eeece599b684820e3f56c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d91c4f461a1912117cb8ed3a8087c5efaaf0a67455b721bfbeaad1c7dda8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5ce3c68997bc5859df929b844d823adf00352454bb278d57e90edcc44817cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5ce3c68997bc5859df929b844d823adf00352454bb278d57e90edcc44817cbe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T04:39:35Z\\\",\\\"message\\\":\\\"l\\\\u003e UUID: UUIDName:} {Op:mutate Table:NB_Global Row:map[] Rows:[] Columns:[] Mutations:[{Column:nb_cfg Mutator:+= Value:1}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {6011affd-30a6-4be6-872d-e4cf1ca780cf}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1129 04:39:35.599606 6479 services_controller.go:434] Service openshift-multus/multus-admission-controller retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{multus-admission-controller openshift-multus c9eea3b1-f918-4c62-9731-c809988317c1 4579 0 2025-02-23 05:21:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:multus-admission-controller] map[service.alpha.openshift.io/serving-cert-secret-name:multus-admission-controller-secret service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{operator.openshift.io/v1 Network cluster 8d01ddba-7e05-4639-926a-4485de3b6327 0xc007916917 0xc007916918}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:webhook,Protocol:TCP,Port:443,TargetPort:{0 6443 },NodePort:0,AppProtocol:nil,},ServicePort{Name:metrics,Protocol:TCP,Port:8443,TargetPort:{1 0 https},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: multus-admission-controller,},ClusterIP:10.217.5.119,Type:ClusterIP,ExternalIPs:[],Sessi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fht8l_openshift-ovn-kubernetes(bd007c61-4689-4179-95ad-44a9724e8e7e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d135e42b8886f1f118abf8f4ce4cf2875ac4dc6c148bfc0677cef6b3651562c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fht8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:40Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.598386 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-qx9wn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce8ce318-27e6-434d-96c8-591d250262fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c99mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c99mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-qx9wn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:40Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.610682 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22ed28a-3748-4eb9-b3a7-c02aa38f7e94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33c0540ccb097b15f65369ff537aef4ae621cc828ba314eb61f0a0037608cadf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c241d29297a73dce12bd0bc2863c56475eb35b8fe8c93991ba0ef8695e38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f42a55d797c6c94af3a5eac1bf08aa38a82359b7ddd0ba32dcc797218ceee6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebed9ec415bb433935da36df8014f9087236cea9788b2048cbff8714f8ad0df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:40Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.622655 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:40Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.627213 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.627247 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.627259 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.627272 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.627280 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:40Z","lastTransitionTime":"2025-11-29T04:39:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.633053 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7lb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55156232-4c80-45a7-aad5-a5fa2793e4ed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6e5473c80737a5b7b00e5874062e0485ed7a4d74632202090a9b12f331953b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d2dtw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7lb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:40Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.647497 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0847696a7b7926047dc62d75739b3b6a7a04cd58efd1c1c2c992d5ae8d20ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b2ec8f0d384066ba90eb42c1a0a925ade3722b2515b0592d6f2a8f5e533a0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-26t88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:40Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.728758 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.728812 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.728824 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.728837 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.728848 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:40Z","lastTransitionTime":"2025-11-29T04:39:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.831634 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.831940 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.832016 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.832105 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.832175 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:40Z","lastTransitionTime":"2025-11-29T04:39:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.934613 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.934644 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.934655 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.934671 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:40 crc kubenswrapper[4799]: I1129 04:39:40.934681 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:40Z","lastTransitionTime":"2025-11-29T04:39:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.037438 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.037472 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.037481 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.037494 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.037502 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:41Z","lastTransitionTime":"2025-11-29T04:39:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.140355 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.140414 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.140430 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.140450 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.140465 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:41Z","lastTransitionTime":"2025-11-29T04:39:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.243545 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.243589 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.243600 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.243619 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.243631 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:41Z","lastTransitionTime":"2025-11-29T04:39:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.348145 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.348246 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.348262 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.348282 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.348296 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:41Z","lastTransitionTime":"2025-11-29T04:39:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.451191 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.451230 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.451238 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.451250 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.451262 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:41Z","lastTransitionTime":"2025-11-29T04:39:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.553861 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.553938 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.553964 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.553994 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.554012 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:41Z","lastTransitionTime":"2025-11-29T04:39:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.659728 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.659889 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qx9wn" Nov 29 04:39:41 crc kubenswrapper[4799]: E1129 04:39:41.660115 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.660197 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.660144 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.660395 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.660423 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:41 crc kubenswrapper[4799]: E1129 04:39:41.660386 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qx9wn" podUID="ce8ce318-27e6-434d-96c8-591d250262fd" Nov 29 04:39:41 crc kubenswrapper[4799]: E1129 04:39:41.660506 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.660473 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.660584 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:41 crc kubenswrapper[4799]: E1129 04:39:41.660587 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.660604 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:41Z","lastTransitionTime":"2025-11-29T04:39:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.763141 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.763224 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.763243 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.763271 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.763289 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:41Z","lastTransitionTime":"2025-11-29T04:39:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.866858 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.866905 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.866913 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.866927 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.866936 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:41Z","lastTransitionTime":"2025-11-29T04:39:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.995698 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.995725 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.995733 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.995747 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:41 crc kubenswrapper[4799]: I1129 04:39:41.995756 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:41Z","lastTransitionTime":"2025-11-29T04:39:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:42 crc kubenswrapper[4799]: I1129 04:39:42.097233 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:42 crc kubenswrapper[4799]: I1129 04:39:42.097610 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:42 crc kubenswrapper[4799]: I1129 04:39:42.097866 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:42 crc kubenswrapper[4799]: I1129 04:39:42.098069 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:42 crc kubenswrapper[4799]: I1129 04:39:42.098208 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:42Z","lastTransitionTime":"2025-11-29T04:39:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:42 crc kubenswrapper[4799]: I1129 04:39:42.200494 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:42 crc kubenswrapper[4799]: I1129 04:39:42.200760 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:42 crc kubenswrapper[4799]: I1129 04:39:42.200935 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:42 crc kubenswrapper[4799]: I1129 04:39:42.201118 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:42 crc kubenswrapper[4799]: I1129 04:39:42.201252 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:42Z","lastTransitionTime":"2025-11-29T04:39:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:42 crc kubenswrapper[4799]: I1129 04:39:42.303483 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:42 crc kubenswrapper[4799]: I1129 04:39:42.303874 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:42 crc kubenswrapper[4799]: I1129 04:39:42.304038 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:42 crc kubenswrapper[4799]: I1129 04:39:42.304182 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:42 crc kubenswrapper[4799]: I1129 04:39:42.304309 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:42Z","lastTransitionTime":"2025-11-29T04:39:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:42 crc kubenswrapper[4799]: I1129 04:39:42.407270 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:42 crc kubenswrapper[4799]: I1129 04:39:42.407316 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:42 crc kubenswrapper[4799]: I1129 04:39:42.407324 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:42 crc kubenswrapper[4799]: I1129 04:39:42.407338 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:42 crc kubenswrapper[4799]: I1129 04:39:42.407347 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:42Z","lastTransitionTime":"2025-11-29T04:39:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:42 crc kubenswrapper[4799]: I1129 04:39:42.510343 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:42 crc kubenswrapper[4799]: I1129 04:39:42.510384 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:42 crc kubenswrapper[4799]: I1129 04:39:42.510396 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:42 crc kubenswrapper[4799]: I1129 04:39:42.510410 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:42 crc kubenswrapper[4799]: I1129 04:39:42.510421 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:42Z","lastTransitionTime":"2025-11-29T04:39:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:42 crc kubenswrapper[4799]: I1129 04:39:42.612293 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:42 crc kubenswrapper[4799]: I1129 04:39:42.612347 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:42 crc kubenswrapper[4799]: I1129 04:39:42.612361 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:42 crc kubenswrapper[4799]: I1129 04:39:42.612380 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:42 crc kubenswrapper[4799]: I1129 04:39:42.612392 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:42Z","lastTransitionTime":"2025-11-29T04:39:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:42 crc kubenswrapper[4799]: I1129 04:39:42.714240 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:42 crc kubenswrapper[4799]: I1129 04:39:42.714292 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:42 crc kubenswrapper[4799]: I1129 04:39:42.714300 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:42 crc kubenswrapper[4799]: I1129 04:39:42.714314 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:42 crc kubenswrapper[4799]: I1129 04:39:42.714356 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:42Z","lastTransitionTime":"2025-11-29T04:39:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:42 crc kubenswrapper[4799]: I1129 04:39:42.816115 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:42 crc kubenswrapper[4799]: I1129 04:39:42.816141 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:42 crc kubenswrapper[4799]: I1129 04:39:42.816148 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:42 crc kubenswrapper[4799]: I1129 04:39:42.816161 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:42 crc kubenswrapper[4799]: I1129 04:39:42.816169 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:42Z","lastTransitionTime":"2025-11-29T04:39:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:42 crc kubenswrapper[4799]: I1129 04:39:42.918239 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:42 crc kubenswrapper[4799]: I1129 04:39:42.918283 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:42 crc kubenswrapper[4799]: I1129 04:39:42.918294 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:42 crc kubenswrapper[4799]: I1129 04:39:42.918310 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:42 crc kubenswrapper[4799]: I1129 04:39:42.918319 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:42Z","lastTransitionTime":"2025-11-29T04:39:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.020382 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.020460 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.020480 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.020506 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.020528 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:43Z","lastTransitionTime":"2025-11-29T04:39:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.122841 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.122916 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.122935 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.122962 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.122979 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:43Z","lastTransitionTime":"2025-11-29T04:39:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.225414 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.225459 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.225471 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.225487 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.225500 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:43Z","lastTransitionTime":"2025-11-29T04:39:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.328010 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.328239 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.328344 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.328419 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.328487 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:43Z","lastTransitionTime":"2025-11-29T04:39:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.430371 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.430453 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.430476 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.430509 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.430533 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:43Z","lastTransitionTime":"2025-11-29T04:39:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.533193 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.533232 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.533240 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.533253 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.533263 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:43Z","lastTransitionTime":"2025-11-29T04:39:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.637262 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.637329 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.637352 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.637385 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.637410 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:43Z","lastTransitionTime":"2025-11-29T04:39:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.658451 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.658532 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.658539 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 04:39:43 crc kubenswrapper[4799]: E1129 04:39:43.658597 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.658640 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qx9wn" Nov 29 04:39:43 crc kubenswrapper[4799]: E1129 04:39:43.658723 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 04:39:43 crc kubenswrapper[4799]: E1129 04:39:43.658811 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 04:39:43 crc kubenswrapper[4799]: E1129 04:39:43.658896 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qx9wn" podUID="ce8ce318-27e6-434d-96c8-591d250262fd" Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.742739 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.742802 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.742813 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.742830 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.742843 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:43Z","lastTransitionTime":"2025-11-29T04:39:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.845944 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.845974 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.845981 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.846024 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.846033 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:43Z","lastTransitionTime":"2025-11-29T04:39:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.947961 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.948002 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.948013 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.948028 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:43 crc kubenswrapper[4799]: I1129 04:39:43.948038 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:43Z","lastTransitionTime":"2025-11-29T04:39:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.050346 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.050406 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.050418 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.050432 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.050442 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:44Z","lastTransitionTime":"2025-11-29T04:39:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.152488 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.152534 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.152546 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.152564 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.152577 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:44Z","lastTransitionTime":"2025-11-29T04:39:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.254740 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.254831 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.254855 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.254884 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.254906 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:44Z","lastTransitionTime":"2025-11-29T04:39:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.357107 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.357168 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.357179 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.357196 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.357209 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:44Z","lastTransitionTime":"2025-11-29T04:39:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.459283 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.459347 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.459360 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.459375 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.459389 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:44Z","lastTransitionTime":"2025-11-29T04:39:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.561918 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.561997 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.562034 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.562054 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.562065 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:44Z","lastTransitionTime":"2025-11-29T04:39:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.664538 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.664587 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.664598 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.664616 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.664631 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:44Z","lastTransitionTime":"2025-11-29T04:39:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.676657 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d40160e-075f-44b7-9d9d-3291e7774f27\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7079548fe9529583a1c129f57838c24cf06d1cb55a97c05c63a05d0cc2e7ec9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7762237003a5a256ff20a16a4869462e2e0dd1426aa391f7ec74ed9b1f572d4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1129 04:39:02.908489 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1129 04:39:02.908637 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 04:39:02.909677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703886220/tls.crt::/tmp/serving-cert-1703886220/tls.key\\\\\\\"\\\\nI1129 04:39:03.077876 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 04:39:03.080368 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 04:39:03.080390 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 04:39:03.080424 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 04:39:03.080432 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 04:39:03.086120 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 04:39:03.086240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086412 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 04:39:03.086502 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 04:39:03.086563 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 04:39:03.086622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1129 04:39:03.086141 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1129 04:39:03.086836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:38:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:44Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.695595 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://396c1c8cbd638c7dc4d8eebded6a17449d9abdc049161578581f1d8c3e31d9a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:44Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.708741 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v79sj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8cf268-e9fd-460a-b579-6411942a00a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c4e099a3149affd1f9f9c5cbaa0f2758d6ab92117c7352db8ede170153c6bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f943aedb7ef6351aa6b2a46b3e84b6db2fae1145bd40f3b9aee6d897aa0d5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v79sj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:44Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.721686 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82a40839-9eab-4511-8494-ee3d2c6ba0b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://186c216d29cdc56583cd5f39acc7cfd0558f60c668b1f5cbf039b566c1747e12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02c75c8691748a626cbef2de964cb1f4579e702276e8b02d805db0876a8fee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://624c4a09beeecd568d6045a720dfa28c15821d931021cedd8776da4bb36b557f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5b8064433d389b05e3e4744ef4c1f72512db615eae94e2a847bbd4cb2518f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d5b8064433d389b05e3e4744ef4c1f72512db615eae94e2a847bbd4cb2518f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:38:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:44Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.737926 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:44Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.759726 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd007c61-4689-4179-95ad-44a9724e8e7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28499112c33332c23be39a11cdf7fc4e1946d04524a3f7c534087179a6620121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dac1c084955d874741d36709fa848c2c475a2d7108d4378b677bb13a88184105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72802532a47b381ead2507e1ae31c49852f9b1d332f3c08fb1e3a972a128eead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c81d94f098efec48c1ff8e172e476bf806ccec5777974eea9c0f3850e5fd40e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3d7f1ba1a408c54f0470fca40e307b70ba208dc7eeece599b684820e3f56c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d91c4f461a1912117cb8ed3a8087c5efaaf0a67455b721bfbeaad1c7dda8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5ce3c68997bc5859df929b844d823adf00352454bb278d57e90edcc44817cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5ce3c68997bc5859df929b844d823adf00352454bb278d57e90edcc44817cbe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T04:39:35Z\\\",\\\"message\\\":\\\"l\\\\u003e UUID: UUIDName:} {Op:mutate Table:NB_Global Row:map[] Rows:[] Columns:[] Mutations:[{Column:nb_cfg Mutator:+= Value:1}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {6011affd-30a6-4be6-872d-e4cf1ca780cf}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1129 04:39:35.599606 6479 services_controller.go:434] Service openshift-multus/multus-admission-controller retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{multus-admission-controller openshift-multus c9eea3b1-f918-4c62-9731-c809988317c1 4579 0 2025-02-23 05:21:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:multus-admission-controller] map[service.alpha.openshift.io/serving-cert-secret-name:multus-admission-controller-secret service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{operator.openshift.io/v1 Network cluster 8d01ddba-7e05-4639-926a-4485de3b6327 0xc007916917 0xc007916918}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:webhook,Protocol:TCP,Port:443,TargetPort:{0 6443 },NodePort:0,AppProtocol:nil,},ServicePort{Name:metrics,Protocol:TCP,Port:8443,TargetPort:{1 0 https},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: multus-admission-controller,},ClusterIP:10.217.5.119,Type:ClusterIP,ExternalIPs:[],Sessi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fht8l_openshift-ovn-kubernetes(bd007c61-4689-4179-95ad-44a9724e8e7e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d135e42b8886f1f118abf8f4ce4cf2875ac4dc6c148bfc0677cef6b3651562c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fht8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:44Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.768355 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.768405 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.768414 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.768429 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.768438 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:44Z","lastTransitionTime":"2025-11-29T04:39:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.773004 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-qx9wn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce8ce318-27e6-434d-96c8-591d250262fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c99mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c99mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-qx9wn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:44Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.791966 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22ed28a-3748-4eb9-b3a7-c02aa38f7e94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33c0540ccb097b15f65369ff537aef4ae621cc828ba314eb61f0a0037608cadf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c241d29297a73dce12bd0bc2863c56475eb35b8fe8c93991ba0ef8695e38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f42a55d797c6c94af3a5eac1bf08aa38a82359b7ddd0ba32dcc797218ceee6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebed9ec415bb433935da36df8014f9087236cea9788b2048cbff8714f8ad0df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:44Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.805562 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:44Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.819306 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7lb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55156232-4c80-45a7-aad5-a5fa2793e4ed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6e5473c80737a5b7b00e5874062e0485ed7a4d74632202090a9b12f331953b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d2dtw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7lb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:44Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.834830 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0847696a7b7926047dc62d75739b3b6a7a04cd58efd1c1c2c992d5ae8d20ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b2ec8f0d384066ba90eb42c1a0a925ade3722b2515b0592d6f2a8f5e533a0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-26t88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:44Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.857838 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s5d7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0af3d308-36b9-4622-8544-635f7b24d687\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23c8830e316099ef7b0e0044145a128658456871418c8c53bcc71210be290df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0088b60069e830e9c36075b09d1ff135da79e94fe6cd271d0e93415af50557e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0088b60069e830e9c36075b09d1ff135da79e94fe6cd271d0e93415af50557e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9b0b72fbe8b552d1266c310f6a6a67822c4c70dec1f812d31cc00748d50565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9b0b72fbe8b552d1266c310f6a6a67822c4c70dec1f812d31cc00748d50565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa3b85ae7c47440f0f61436e1cb1f3f2096e0e4011434876f9d17e913ca24df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa3b85ae7c47440f0f61436e1cb1f3f2096e0e4011434876f9d17e913ca24df7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s5d7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:44Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.872506 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.872570 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.872586 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.872610 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.872627 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:44Z","lastTransitionTime":"2025-11-29T04:39:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.873206 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d87rr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a92ae69-5675-49d9-a567-e2476fe02ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://207d152294f9996ed7294d858842da5eef891a6d8a027473dc3d29a7101aadab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffwm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d87rr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:44Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.894033 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://deb9f370792598608cf2e0c0fe9975eaf8b95dc94b12eef8c55112df4aa9c998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:44Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.910241 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:44Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.923338 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee17b4a9a8999e5f76ceaf69dbd13acf80ed0e95e72f3e1428ce38c5d5982788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://382256e0fa1a4c26610227b9ae96288ccbfbbd97aa3cbd5fd4f18cb4228a234b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:44Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.935537 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt2dl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6c9236-5f7d-46c5-b080-4667620d874e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b6c827d0c57a87808fef301661fce28c50359e85c9e47947c50945f4f869de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt2dl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:44Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.974412 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.974470 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.974487 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.974510 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:44 crc kubenswrapper[4799]: I1129 04:39:44.974700 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:44Z","lastTransitionTime":"2025-11-29T04:39:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:45 crc kubenswrapper[4799]: I1129 04:39:45.076467 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:45 crc kubenswrapper[4799]: I1129 04:39:45.076509 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:45 crc kubenswrapper[4799]: I1129 04:39:45.076517 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:45 crc kubenswrapper[4799]: I1129 04:39:45.076530 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:45 crc kubenswrapper[4799]: I1129 04:39:45.076539 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:45Z","lastTransitionTime":"2025-11-29T04:39:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:45 crc kubenswrapper[4799]: I1129 04:39:45.179648 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:45 crc kubenswrapper[4799]: I1129 04:39:45.179699 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:45 crc kubenswrapper[4799]: I1129 04:39:45.179709 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:45 crc kubenswrapper[4799]: I1129 04:39:45.179723 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:45 crc kubenswrapper[4799]: I1129 04:39:45.179733 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:45Z","lastTransitionTime":"2025-11-29T04:39:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:45 crc kubenswrapper[4799]: I1129 04:39:45.282802 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:45 crc kubenswrapper[4799]: I1129 04:39:45.283012 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:45 crc kubenswrapper[4799]: I1129 04:39:45.283022 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:45 crc kubenswrapper[4799]: I1129 04:39:45.283036 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:45 crc kubenswrapper[4799]: I1129 04:39:45.283044 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:45Z","lastTransitionTime":"2025-11-29T04:39:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:45 crc kubenswrapper[4799]: I1129 04:39:45.385836 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:45 crc kubenswrapper[4799]: I1129 04:39:45.385882 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:45 crc kubenswrapper[4799]: I1129 04:39:45.385893 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:45 crc kubenswrapper[4799]: I1129 04:39:45.385908 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:45 crc kubenswrapper[4799]: I1129 04:39:45.385920 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:45Z","lastTransitionTime":"2025-11-29T04:39:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:45 crc kubenswrapper[4799]: I1129 04:39:45.487693 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:45 crc kubenswrapper[4799]: I1129 04:39:45.487728 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:45 crc kubenswrapper[4799]: I1129 04:39:45.487736 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:45 crc kubenswrapper[4799]: I1129 04:39:45.487748 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:45 crc kubenswrapper[4799]: I1129 04:39:45.487757 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:45Z","lastTransitionTime":"2025-11-29T04:39:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:45 crc kubenswrapper[4799]: I1129 04:39:45.591651 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:45 crc kubenswrapper[4799]: I1129 04:39:45.591728 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:45 crc kubenswrapper[4799]: I1129 04:39:45.591747 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:45 crc kubenswrapper[4799]: I1129 04:39:45.591778 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:45 crc kubenswrapper[4799]: I1129 04:39:45.591839 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:45Z","lastTransitionTime":"2025-11-29T04:39:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:45 crc kubenswrapper[4799]: I1129 04:39:45.659194 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:39:45 crc kubenswrapper[4799]: E1129 04:39:45.660338 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 04:39:45 crc kubenswrapper[4799]: I1129 04:39:45.660526 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 04:39:45 crc kubenswrapper[4799]: I1129 04:39:45.660536 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 04:39:45 crc kubenswrapper[4799]: E1129 04:39:45.660715 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 04:39:45 crc kubenswrapper[4799]: I1129 04:39:45.660740 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qx9wn" Nov 29 04:39:45 crc kubenswrapper[4799]: E1129 04:39:45.660968 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 04:39:45 crc kubenswrapper[4799]: E1129 04:39:45.661169 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qx9wn" podUID="ce8ce318-27e6-434d-96c8-591d250262fd" Nov 29 04:39:45 crc kubenswrapper[4799]: I1129 04:39:45.693727 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:45 crc kubenswrapper[4799]: I1129 04:39:45.693761 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:45 crc kubenswrapper[4799]: I1129 04:39:45.693769 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:45 crc kubenswrapper[4799]: I1129 04:39:45.693782 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:45 crc kubenswrapper[4799]: I1129 04:39:45.693816 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:45Z","lastTransitionTime":"2025-11-29T04:39:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:45 crc kubenswrapper[4799]: I1129 04:39:45.796491 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:45 crc kubenswrapper[4799]: I1129 04:39:45.796533 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:45 crc kubenswrapper[4799]: I1129 04:39:45.796541 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:45 crc kubenswrapper[4799]: I1129 04:39:45.796558 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:45 crc kubenswrapper[4799]: I1129 04:39:45.796568 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:45Z","lastTransitionTime":"2025-11-29T04:39:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:45 crc kubenswrapper[4799]: I1129 04:39:45.899317 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:45 crc kubenswrapper[4799]: I1129 04:39:45.899370 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:45 crc kubenswrapper[4799]: I1129 04:39:45.899380 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:45 crc kubenswrapper[4799]: I1129 04:39:45.899396 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:45 crc kubenswrapper[4799]: I1129 04:39:45.899408 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:45Z","lastTransitionTime":"2025-11-29T04:39:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:46 crc kubenswrapper[4799]: I1129 04:39:46.004045 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:46 crc kubenswrapper[4799]: I1129 04:39:46.004094 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:46 crc kubenswrapper[4799]: I1129 04:39:46.004107 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:46 crc kubenswrapper[4799]: I1129 04:39:46.004127 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:46 crc kubenswrapper[4799]: I1129 04:39:46.004139 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:46Z","lastTransitionTime":"2025-11-29T04:39:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:46 crc kubenswrapper[4799]: I1129 04:39:46.106812 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:46 crc kubenswrapper[4799]: I1129 04:39:46.107359 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:46 crc kubenswrapper[4799]: I1129 04:39:46.107478 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:46 crc kubenswrapper[4799]: I1129 04:39:46.107836 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:46 crc kubenswrapper[4799]: I1129 04:39:46.107925 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:46Z","lastTransitionTime":"2025-11-29T04:39:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:46 crc kubenswrapper[4799]: I1129 04:39:46.212305 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:46 crc kubenswrapper[4799]: I1129 04:39:46.212386 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:46 crc kubenswrapper[4799]: I1129 04:39:46.212407 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:46 crc kubenswrapper[4799]: I1129 04:39:46.212449 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:46 crc kubenswrapper[4799]: I1129 04:39:46.212470 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:46Z","lastTransitionTime":"2025-11-29T04:39:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:46 crc kubenswrapper[4799]: I1129 04:39:46.316621 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:46 crc kubenswrapper[4799]: I1129 04:39:46.316700 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:46 crc kubenswrapper[4799]: I1129 04:39:46.316722 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:46 crc kubenswrapper[4799]: I1129 04:39:46.316751 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:46 crc kubenswrapper[4799]: I1129 04:39:46.316771 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:46Z","lastTransitionTime":"2025-11-29T04:39:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:46 crc kubenswrapper[4799]: I1129 04:39:46.421135 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:46 crc kubenswrapper[4799]: I1129 04:39:46.421181 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:46 crc kubenswrapper[4799]: I1129 04:39:46.421191 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:46 crc kubenswrapper[4799]: I1129 04:39:46.421368 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:46 crc kubenswrapper[4799]: I1129 04:39:46.421382 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:46Z","lastTransitionTime":"2025-11-29T04:39:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:46 crc kubenswrapper[4799]: I1129 04:39:46.524649 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:46 crc kubenswrapper[4799]: I1129 04:39:46.524703 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:46 crc kubenswrapper[4799]: I1129 04:39:46.524721 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:46 crc kubenswrapper[4799]: I1129 04:39:46.524748 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:46 crc kubenswrapper[4799]: I1129 04:39:46.524766 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:46Z","lastTransitionTime":"2025-11-29T04:39:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:46 crc kubenswrapper[4799]: I1129 04:39:46.627807 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:46 crc kubenswrapper[4799]: I1129 04:39:46.627849 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:46 crc kubenswrapper[4799]: I1129 04:39:46.627858 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:46 crc kubenswrapper[4799]: I1129 04:39:46.627873 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:46 crc kubenswrapper[4799]: I1129 04:39:46.627885 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:46Z","lastTransitionTime":"2025-11-29T04:39:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:46 crc kubenswrapper[4799]: I1129 04:39:46.730210 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:46 crc kubenswrapper[4799]: I1129 04:39:46.730247 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:46 crc kubenswrapper[4799]: I1129 04:39:46.730261 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:46 crc kubenswrapper[4799]: I1129 04:39:46.730276 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:46 crc kubenswrapper[4799]: I1129 04:39:46.730285 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:46Z","lastTransitionTime":"2025-11-29T04:39:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:46 crc kubenswrapper[4799]: I1129 04:39:46.833187 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:46 crc kubenswrapper[4799]: I1129 04:39:46.833237 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:46 crc kubenswrapper[4799]: I1129 04:39:46.833255 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:46 crc kubenswrapper[4799]: I1129 04:39:46.833277 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:46 crc kubenswrapper[4799]: I1129 04:39:46.833293 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:46Z","lastTransitionTime":"2025-11-29T04:39:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:46 crc kubenswrapper[4799]: I1129 04:39:46.935397 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:46 crc kubenswrapper[4799]: I1129 04:39:46.935530 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:46 crc kubenswrapper[4799]: I1129 04:39:46.935557 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:46 crc kubenswrapper[4799]: I1129 04:39:46.935589 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:46 crc kubenswrapper[4799]: I1129 04:39:46.935615 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:46Z","lastTransitionTime":"2025-11-29T04:39:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.038933 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.039016 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.039049 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.039100 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.039123 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:47Z","lastTransitionTime":"2025-11-29T04:39:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.142734 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.143052 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.143192 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.143267 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.143323 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:47Z","lastTransitionTime":"2025-11-29T04:39:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.245977 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.246215 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.246330 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.246407 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.246473 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:47Z","lastTransitionTime":"2025-11-29T04:39:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.348592 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.348645 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.348660 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.348680 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.348698 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:47Z","lastTransitionTime":"2025-11-29T04:39:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.451696 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.451741 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.451755 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.451783 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.451820 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:47Z","lastTransitionTime":"2025-11-29T04:39:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.554114 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.554154 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.554165 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.554192 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.554204 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:47Z","lastTransitionTime":"2025-11-29T04:39:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.656325 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.656396 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.656413 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.656439 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.656464 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:47Z","lastTransitionTime":"2025-11-29T04:39:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.658502 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qx9wn" Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.658591 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.658599 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:39:47 crc kubenswrapper[4799]: E1129 04:39:47.658678 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qx9wn" podUID="ce8ce318-27e6-434d-96c8-591d250262fd" Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.658733 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 04:39:47 crc kubenswrapper[4799]: E1129 04:39:47.658855 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 04:39:47 crc kubenswrapper[4799]: E1129 04:39:47.659004 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 04:39:47 crc kubenswrapper[4799]: E1129 04:39:47.659177 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.758867 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.758945 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.758957 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.758977 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.758989 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:47Z","lastTransitionTime":"2025-11-29T04:39:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.861336 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.861402 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.861422 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.861445 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.861463 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:47Z","lastTransitionTime":"2025-11-29T04:39:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.963826 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.963866 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.963877 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.963893 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:47 crc kubenswrapper[4799]: I1129 04:39:47.963904 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:47Z","lastTransitionTime":"2025-11-29T04:39:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.066281 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.066334 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.066346 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.066367 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.066379 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:48Z","lastTransitionTime":"2025-11-29T04:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.169409 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.169449 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.169461 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.169476 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.169489 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:48Z","lastTransitionTime":"2025-11-29T04:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.271878 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.271948 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.271971 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.272004 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.272028 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:48Z","lastTransitionTime":"2025-11-29T04:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.382718 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.382763 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.382773 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.382808 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.382819 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:48Z","lastTransitionTime":"2025-11-29T04:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.485156 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.485207 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.485220 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.485238 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.485251 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:48Z","lastTransitionTime":"2025-11-29T04:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.587987 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.588052 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.588071 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.588124 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.588142 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:48Z","lastTransitionTime":"2025-11-29T04:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.686086 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.686141 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.686153 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.686171 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.686183 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:48Z","lastTransitionTime":"2025-11-29T04:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:48 crc kubenswrapper[4799]: E1129 04:39:48.701354 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6f4e4672-da32-440c-8c4f-6e7f4942f55e\\\",\\\"systemUUID\\\":\\\"882c4d8e-57b7-4312-8499-40d1f85d56c9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:48Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.705480 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.705509 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.705523 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.705540 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.705553 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:48Z","lastTransitionTime":"2025-11-29T04:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:48 crc kubenswrapper[4799]: E1129 04:39:48.720122 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6f4e4672-da32-440c-8c4f-6e7f4942f55e\\\",\\\"systemUUID\\\":\\\"882c4d8e-57b7-4312-8499-40d1f85d56c9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:48Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.724051 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.724081 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.724093 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.724125 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.724137 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:48Z","lastTransitionTime":"2025-11-29T04:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:48 crc kubenswrapper[4799]: E1129 04:39:48.737866 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6f4e4672-da32-440c-8c4f-6e7f4942f55e\\\",\\\"systemUUID\\\":\\\"882c4d8e-57b7-4312-8499-40d1f85d56c9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:48Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.742040 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.742073 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.742084 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.742107 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.742118 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:48Z","lastTransitionTime":"2025-11-29T04:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:48 crc kubenswrapper[4799]: E1129 04:39:48.754534 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6f4e4672-da32-440c-8c4f-6e7f4942f55e\\\",\\\"systemUUID\\\":\\\"882c4d8e-57b7-4312-8499-40d1f85d56c9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:48Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.758559 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.758615 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.758681 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.758725 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.758744 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:48Z","lastTransitionTime":"2025-11-29T04:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:48 crc kubenswrapper[4799]: E1129 04:39:48.772278 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6f4e4672-da32-440c-8c4f-6e7f4942f55e\\\",\\\"systemUUID\\\":\\\"882c4d8e-57b7-4312-8499-40d1f85d56c9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:48Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:48 crc kubenswrapper[4799]: E1129 04:39:48.772470 4799 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.774319 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.774345 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.774358 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.774390 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.774402 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:48Z","lastTransitionTime":"2025-11-29T04:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.876484 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.876539 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.876549 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.876589 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.876602 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:48Z","lastTransitionTime":"2025-11-29T04:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.980456 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.980513 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.980529 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.980563 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:48 crc kubenswrapper[4799]: I1129 04:39:48.980579 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:48Z","lastTransitionTime":"2025-11-29T04:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:49 crc kubenswrapper[4799]: I1129 04:39:49.084053 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:49 crc kubenswrapper[4799]: I1129 04:39:49.084098 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:49 crc kubenswrapper[4799]: I1129 04:39:49.084108 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:49 crc kubenswrapper[4799]: I1129 04:39:49.084123 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:49 crc kubenswrapper[4799]: I1129 04:39:49.084135 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:49Z","lastTransitionTime":"2025-11-29T04:39:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:49 crc kubenswrapper[4799]: I1129 04:39:49.188380 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:49 crc kubenswrapper[4799]: I1129 04:39:49.188411 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:49 crc kubenswrapper[4799]: I1129 04:39:49.188421 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:49 crc kubenswrapper[4799]: I1129 04:39:49.188434 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:49 crc kubenswrapper[4799]: I1129 04:39:49.188443 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:49Z","lastTransitionTime":"2025-11-29T04:39:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:49 crc kubenswrapper[4799]: I1129 04:39:49.291860 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:49 crc kubenswrapper[4799]: I1129 04:39:49.291913 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:49 crc kubenswrapper[4799]: I1129 04:39:49.291926 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:49 crc kubenswrapper[4799]: I1129 04:39:49.291948 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:49 crc kubenswrapper[4799]: I1129 04:39:49.291958 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:49Z","lastTransitionTime":"2025-11-29T04:39:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:49 crc kubenswrapper[4799]: I1129 04:39:49.395247 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:49 crc kubenswrapper[4799]: I1129 04:39:49.395325 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:49 crc kubenswrapper[4799]: I1129 04:39:49.395349 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:49 crc kubenswrapper[4799]: I1129 04:39:49.395381 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:49 crc kubenswrapper[4799]: I1129 04:39:49.395401 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:49Z","lastTransitionTime":"2025-11-29T04:39:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:49 crc kubenswrapper[4799]: I1129 04:39:49.500398 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:49 crc kubenswrapper[4799]: I1129 04:39:49.500481 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:49 crc kubenswrapper[4799]: I1129 04:39:49.500507 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:49 crc kubenswrapper[4799]: I1129 04:39:49.500555 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:49 crc kubenswrapper[4799]: I1129 04:39:49.500584 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:49Z","lastTransitionTime":"2025-11-29T04:39:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:49 crc kubenswrapper[4799]: I1129 04:39:49.603712 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:49 crc kubenswrapper[4799]: I1129 04:39:49.603773 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:49 crc kubenswrapper[4799]: I1129 04:39:49.603835 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:49 crc kubenswrapper[4799]: I1129 04:39:49.603866 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:49 crc kubenswrapper[4799]: I1129 04:39:49.603887 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:49Z","lastTransitionTime":"2025-11-29T04:39:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:49 crc kubenswrapper[4799]: I1129 04:39:49.658083 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 04:39:49 crc kubenswrapper[4799]: I1129 04:39:49.658187 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 04:39:49 crc kubenswrapper[4799]: I1129 04:39:49.658116 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:39:49 crc kubenswrapper[4799]: I1129 04:39:49.658238 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qx9wn" Nov 29 04:39:49 crc kubenswrapper[4799]: E1129 04:39:49.658301 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 04:39:49 crc kubenswrapper[4799]: E1129 04:39:49.658431 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 04:39:49 crc kubenswrapper[4799]: E1129 04:39:49.658600 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qx9wn" podUID="ce8ce318-27e6-434d-96c8-591d250262fd" Nov 29 04:39:49 crc kubenswrapper[4799]: E1129 04:39:49.658662 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 04:39:49 crc kubenswrapper[4799]: I1129 04:39:49.708081 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:49 crc kubenswrapper[4799]: I1129 04:39:49.708167 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:49 crc kubenswrapper[4799]: I1129 04:39:49.708191 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:49 crc kubenswrapper[4799]: I1129 04:39:49.708223 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:49 crc kubenswrapper[4799]: I1129 04:39:49.708246 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:49Z","lastTransitionTime":"2025-11-29T04:39:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:49 crc kubenswrapper[4799]: I1129 04:39:49.812310 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:49 crc kubenswrapper[4799]: I1129 04:39:49.812374 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:49 crc kubenswrapper[4799]: I1129 04:39:49.812394 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:49 crc kubenswrapper[4799]: I1129 04:39:49.812423 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:49 crc kubenswrapper[4799]: I1129 04:39:49.812444 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:49Z","lastTransitionTime":"2025-11-29T04:39:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:49 crc kubenswrapper[4799]: I1129 04:39:49.916311 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:49 crc kubenswrapper[4799]: I1129 04:39:49.916413 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:49 crc kubenswrapper[4799]: I1129 04:39:49.916432 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:49 crc kubenswrapper[4799]: I1129 04:39:49.916471 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:49 crc kubenswrapper[4799]: I1129 04:39:49.916492 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:49Z","lastTransitionTime":"2025-11-29T04:39:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:50 crc kubenswrapper[4799]: I1129 04:39:50.019932 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:50 crc kubenswrapper[4799]: I1129 04:39:50.019976 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:50 crc kubenswrapper[4799]: I1129 04:39:50.019987 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:50 crc kubenswrapper[4799]: I1129 04:39:50.020006 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:50 crc kubenswrapper[4799]: I1129 04:39:50.020018 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:50Z","lastTransitionTime":"2025-11-29T04:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:50 crc kubenswrapper[4799]: I1129 04:39:50.125263 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:50 crc kubenswrapper[4799]: I1129 04:39:50.125320 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:50 crc kubenswrapper[4799]: I1129 04:39:50.125330 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:50 crc kubenswrapper[4799]: I1129 04:39:50.125349 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:50 crc kubenswrapper[4799]: I1129 04:39:50.125365 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:50Z","lastTransitionTime":"2025-11-29T04:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:50 crc kubenswrapper[4799]: I1129 04:39:50.228405 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:50 crc kubenswrapper[4799]: I1129 04:39:50.228449 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:50 crc kubenswrapper[4799]: I1129 04:39:50.228485 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:50 crc kubenswrapper[4799]: I1129 04:39:50.228508 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:50 crc kubenswrapper[4799]: I1129 04:39:50.228522 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:50Z","lastTransitionTime":"2025-11-29T04:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:50 crc kubenswrapper[4799]: I1129 04:39:50.330644 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:50 crc kubenswrapper[4799]: I1129 04:39:50.330681 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:50 crc kubenswrapper[4799]: I1129 04:39:50.330761 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:50 crc kubenswrapper[4799]: I1129 04:39:50.330780 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:50 crc kubenswrapper[4799]: I1129 04:39:50.330823 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:50Z","lastTransitionTime":"2025-11-29T04:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:50 crc kubenswrapper[4799]: I1129 04:39:50.436470 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:50 crc kubenswrapper[4799]: I1129 04:39:50.436519 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:50 crc kubenswrapper[4799]: I1129 04:39:50.436529 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:50 crc kubenswrapper[4799]: I1129 04:39:50.436547 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:50 crc kubenswrapper[4799]: I1129 04:39:50.436558 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:50Z","lastTransitionTime":"2025-11-29T04:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:50 crc kubenswrapper[4799]: I1129 04:39:50.539655 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:50 crc kubenswrapper[4799]: I1129 04:39:50.540026 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:50 crc kubenswrapper[4799]: I1129 04:39:50.540129 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:50 crc kubenswrapper[4799]: I1129 04:39:50.540224 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:50 crc kubenswrapper[4799]: I1129 04:39:50.540320 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:50Z","lastTransitionTime":"2025-11-29T04:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:50 crc kubenswrapper[4799]: I1129 04:39:50.643303 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:50 crc kubenswrapper[4799]: I1129 04:39:50.643345 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:50 crc kubenswrapper[4799]: I1129 04:39:50.643356 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:50 crc kubenswrapper[4799]: I1129 04:39:50.643373 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:50 crc kubenswrapper[4799]: I1129 04:39:50.643384 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:50Z","lastTransitionTime":"2025-11-29T04:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:50 crc kubenswrapper[4799]: I1129 04:39:50.746719 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:50 crc kubenswrapper[4799]: I1129 04:39:50.747157 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:50 crc kubenswrapper[4799]: I1129 04:39:50.747223 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:50 crc kubenswrapper[4799]: I1129 04:39:50.747310 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:50 crc kubenswrapper[4799]: I1129 04:39:50.747372 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:50Z","lastTransitionTime":"2025-11-29T04:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:50 crc kubenswrapper[4799]: I1129 04:39:50.849504 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:50 crc kubenswrapper[4799]: I1129 04:39:50.849541 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:50 crc kubenswrapper[4799]: I1129 04:39:50.849550 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:50 crc kubenswrapper[4799]: I1129 04:39:50.849564 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:50 crc kubenswrapper[4799]: I1129 04:39:50.849573 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:50Z","lastTransitionTime":"2025-11-29T04:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:50 crc kubenswrapper[4799]: I1129 04:39:50.952199 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:50 crc kubenswrapper[4799]: I1129 04:39:50.952235 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:50 crc kubenswrapper[4799]: I1129 04:39:50.952245 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:50 crc kubenswrapper[4799]: I1129 04:39:50.952259 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:50 crc kubenswrapper[4799]: I1129 04:39:50.952269 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:50Z","lastTransitionTime":"2025-11-29T04:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.054772 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.054852 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.054865 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.054887 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.054902 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:51Z","lastTransitionTime":"2025-11-29T04:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.157844 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.157902 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.157914 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.157935 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.157948 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:51Z","lastTransitionTime":"2025-11-29T04:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.259823 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.259862 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.259873 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.259885 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.259895 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:51Z","lastTransitionTime":"2025-11-29T04:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.362377 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.362423 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.362434 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.362449 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.362460 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:51Z","lastTransitionTime":"2025-11-29T04:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.464780 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.464842 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.464854 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.464871 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.464883 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:51Z","lastTransitionTime":"2025-11-29T04:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.567041 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.567074 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.567083 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.567098 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.567106 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:51Z","lastTransitionTime":"2025-11-29T04:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.657909 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.657925 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.658016 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.658112 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qx9wn" Nov 29 04:39:51 crc kubenswrapper[4799]: E1129 04:39:51.658228 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 04:39:51 crc kubenswrapper[4799]: E1129 04:39:51.658468 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 04:39:51 crc kubenswrapper[4799]: E1129 04:39:51.658515 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 04:39:51 crc kubenswrapper[4799]: E1129 04:39:51.658631 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qx9wn" podUID="ce8ce318-27e6-434d-96c8-591d250262fd" Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.669616 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.669677 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.669696 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.669721 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.669742 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:51Z","lastTransitionTime":"2025-11-29T04:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.771968 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.772005 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.772014 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.772028 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.772036 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:51Z","lastTransitionTime":"2025-11-29T04:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.875351 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.875404 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.875419 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.875440 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.875452 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:51Z","lastTransitionTime":"2025-11-29T04:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.977921 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.977952 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.977960 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.977973 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:51 crc kubenswrapper[4799]: I1129 04:39:51.977982 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:51Z","lastTransitionTime":"2025-11-29T04:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:52 crc kubenswrapper[4799]: I1129 04:39:52.082157 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:52 crc kubenswrapper[4799]: I1129 04:39:52.082234 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:52 crc kubenswrapper[4799]: I1129 04:39:52.082259 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:52 crc kubenswrapper[4799]: I1129 04:39:52.082295 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:52 crc kubenswrapper[4799]: I1129 04:39:52.082323 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:52Z","lastTransitionTime":"2025-11-29T04:39:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:52 crc kubenswrapper[4799]: I1129 04:39:52.185955 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:52 crc kubenswrapper[4799]: I1129 04:39:52.186047 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:52 crc kubenswrapper[4799]: I1129 04:39:52.186068 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:52 crc kubenswrapper[4799]: I1129 04:39:52.186109 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:52 crc kubenswrapper[4799]: I1129 04:39:52.186130 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:52Z","lastTransitionTime":"2025-11-29T04:39:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:52 crc kubenswrapper[4799]: I1129 04:39:52.288916 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:52 crc kubenswrapper[4799]: I1129 04:39:52.288956 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:52 crc kubenswrapper[4799]: I1129 04:39:52.288965 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:52 crc kubenswrapper[4799]: I1129 04:39:52.288985 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:52 crc kubenswrapper[4799]: I1129 04:39:52.288997 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:52Z","lastTransitionTime":"2025-11-29T04:39:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:52 crc kubenswrapper[4799]: I1129 04:39:52.392532 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:52 crc kubenswrapper[4799]: I1129 04:39:52.392579 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:52 crc kubenswrapper[4799]: I1129 04:39:52.392591 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:52 crc kubenswrapper[4799]: I1129 04:39:52.392610 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:52 crc kubenswrapper[4799]: I1129 04:39:52.392622 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:52Z","lastTransitionTime":"2025-11-29T04:39:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:52 crc kubenswrapper[4799]: I1129 04:39:52.495476 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:52 crc kubenswrapper[4799]: I1129 04:39:52.495559 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:52 crc kubenswrapper[4799]: I1129 04:39:52.495581 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:52 crc kubenswrapper[4799]: I1129 04:39:52.495614 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:52 crc kubenswrapper[4799]: I1129 04:39:52.495635 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:52Z","lastTransitionTime":"2025-11-29T04:39:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:52 crc kubenswrapper[4799]: I1129 04:39:52.598834 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:52 crc kubenswrapper[4799]: I1129 04:39:52.598898 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:52 crc kubenswrapper[4799]: I1129 04:39:52.598915 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:52 crc kubenswrapper[4799]: I1129 04:39:52.598934 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:52 crc kubenswrapper[4799]: I1129 04:39:52.598946 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:52Z","lastTransitionTime":"2025-11-29T04:39:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:52 crc kubenswrapper[4799]: I1129 04:39:52.701217 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:52 crc kubenswrapper[4799]: I1129 04:39:52.701258 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:52 crc kubenswrapper[4799]: I1129 04:39:52.701268 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:52 crc kubenswrapper[4799]: I1129 04:39:52.701286 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:52 crc kubenswrapper[4799]: I1129 04:39:52.701296 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:52Z","lastTransitionTime":"2025-11-29T04:39:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:52 crc kubenswrapper[4799]: I1129 04:39:52.804157 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:52 crc kubenswrapper[4799]: I1129 04:39:52.804197 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:52 crc kubenswrapper[4799]: I1129 04:39:52.804212 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:52 crc kubenswrapper[4799]: I1129 04:39:52.804228 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:52 crc kubenswrapper[4799]: I1129 04:39:52.804239 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:52Z","lastTransitionTime":"2025-11-29T04:39:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:52 crc kubenswrapper[4799]: I1129 04:39:52.906420 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:52 crc kubenswrapper[4799]: I1129 04:39:52.906487 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:52 crc kubenswrapper[4799]: I1129 04:39:52.906506 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:52 crc kubenswrapper[4799]: I1129 04:39:52.906533 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:52 crc kubenswrapper[4799]: I1129 04:39:52.906550 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:52Z","lastTransitionTime":"2025-11-29T04:39:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:53 crc kubenswrapper[4799]: I1129 04:39:53.009064 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:53 crc kubenswrapper[4799]: I1129 04:39:53.009128 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:53 crc kubenswrapper[4799]: I1129 04:39:53.009140 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:53 crc kubenswrapper[4799]: I1129 04:39:53.009156 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:53 crc kubenswrapper[4799]: I1129 04:39:53.009166 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:53Z","lastTransitionTime":"2025-11-29T04:39:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:53 crc kubenswrapper[4799]: I1129 04:39:53.112286 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:53 crc kubenswrapper[4799]: I1129 04:39:53.112519 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:53 crc kubenswrapper[4799]: I1129 04:39:53.112586 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:53 crc kubenswrapper[4799]: I1129 04:39:53.112646 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:53 crc kubenswrapper[4799]: I1129 04:39:53.112708 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:53Z","lastTransitionTime":"2025-11-29T04:39:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:53 crc kubenswrapper[4799]: I1129 04:39:53.215403 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:53 crc kubenswrapper[4799]: I1129 04:39:53.215633 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:53 crc kubenswrapper[4799]: I1129 04:39:53.215699 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:53 crc kubenswrapper[4799]: I1129 04:39:53.215761 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:53 crc kubenswrapper[4799]: I1129 04:39:53.215855 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:53Z","lastTransitionTime":"2025-11-29T04:39:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:53 crc kubenswrapper[4799]: I1129 04:39:53.319257 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:53 crc kubenswrapper[4799]: I1129 04:39:53.319581 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:53 crc kubenswrapper[4799]: I1129 04:39:53.319713 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:53 crc kubenswrapper[4799]: I1129 04:39:53.319894 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:53 crc kubenswrapper[4799]: I1129 04:39:53.320046 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:53Z","lastTransitionTime":"2025-11-29T04:39:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:53 crc kubenswrapper[4799]: I1129 04:39:53.426326 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:53 crc kubenswrapper[4799]: I1129 04:39:53.426893 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:53 crc kubenswrapper[4799]: I1129 04:39:53.427056 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:53 crc kubenswrapper[4799]: I1129 04:39:53.427204 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:53 crc kubenswrapper[4799]: I1129 04:39:53.427352 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:53Z","lastTransitionTime":"2025-11-29T04:39:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:53 crc kubenswrapper[4799]: I1129 04:39:53.534122 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:53 crc kubenswrapper[4799]: I1129 04:39:53.534597 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:53 crc kubenswrapper[4799]: I1129 04:39:53.534785 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:53 crc kubenswrapper[4799]: I1129 04:39:53.534968 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:53 crc kubenswrapper[4799]: I1129 04:39:53.535134 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:53Z","lastTransitionTime":"2025-11-29T04:39:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:53 crc kubenswrapper[4799]: I1129 04:39:53.641031 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:53 crc kubenswrapper[4799]: I1129 04:39:53.641602 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:53 crc kubenswrapper[4799]: I1129 04:39:53.641866 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:53 crc kubenswrapper[4799]: I1129 04:39:53.642077 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:53 crc kubenswrapper[4799]: I1129 04:39:53.642261 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:53Z","lastTransitionTime":"2025-11-29T04:39:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:53 crc kubenswrapper[4799]: I1129 04:39:53.746823 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:53 crc kubenswrapper[4799]: I1129 04:39:53.747518 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:53 crc kubenswrapper[4799]: I1129 04:39:53.747651 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:53 crc kubenswrapper[4799]: I1129 04:39:53.747851 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:53 crc kubenswrapper[4799]: I1129 04:39:53.748011 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:53Z","lastTransitionTime":"2025-11-29T04:39:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:53 crc kubenswrapper[4799]: I1129 04:39:53.852098 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:53 crc kubenswrapper[4799]: I1129 04:39:53.853632 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:53 crc kubenswrapper[4799]: I1129 04:39:53.853697 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:53 crc kubenswrapper[4799]: I1129 04:39:53.853735 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:53 crc kubenswrapper[4799]: I1129 04:39:53.853765 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:53Z","lastTransitionTime":"2025-11-29T04:39:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:53 crc kubenswrapper[4799]: I1129 04:39:53.959349 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:53 crc kubenswrapper[4799]: I1129 04:39:53.959410 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:53 crc kubenswrapper[4799]: I1129 04:39:53.959428 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:53 crc kubenswrapper[4799]: I1129 04:39:53.959455 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:53 crc kubenswrapper[4799]: I1129 04:39:53.959474 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:53Z","lastTransitionTime":"2025-11-29T04:39:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.063678 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.063737 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.063757 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.063785 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.063842 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:54Z","lastTransitionTime":"2025-11-29T04:39:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.191368 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 04:39:54 crc kubenswrapper[4799]: E1129 04:39:54.191563 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.191846 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 04:39:54 crc kubenswrapper[4799]: E1129 04:39:54.192051 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.192445 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:39:54 crc kubenswrapper[4799]: E1129 04:39:54.192535 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.195561 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.195645 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.195661 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.195682 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.195704 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:54Z","lastTransitionTime":"2025-11-29T04:39:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.203971 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qx9wn" Nov 29 04:39:54 crc kubenswrapper[4799]: E1129 04:39:54.204487 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qx9wn" podUID="ce8ce318-27e6-434d-96c8-591d250262fd" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.204897 4799 scope.go:117] "RemoveContainer" containerID="d5ce3c68997bc5859df929b844d823adf00352454bb278d57e90edcc44817cbe" Nov 29 04:39:54 crc kubenswrapper[4799]: E1129 04:39:54.205448 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fht8l_openshift-ovn-kubernetes(bd007c61-4689-4179-95ad-44a9724e8e7e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" podUID="bd007c61-4689-4179-95ad-44a9724e8e7e" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.304950 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.305019 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.305037 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.305065 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.305083 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:54Z","lastTransitionTime":"2025-11-29T04:39:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.408055 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.408114 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.408128 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.408149 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.408529 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:54Z","lastTransitionTime":"2025-11-29T04:39:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.512078 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.512147 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.512171 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.512201 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.512223 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:54Z","lastTransitionTime":"2025-11-29T04:39:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.615665 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.615702 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.615710 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.615723 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.615763 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:54Z","lastTransitionTime":"2025-11-29T04:39:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.674498 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://396c1c8cbd638c7dc4d8eebded6a17449d9abdc049161578581f1d8c3e31d9a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:54Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.685549 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v79sj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8cf268-e9fd-460a-b579-6411942a00a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c4e099a3149affd1f9f9c5cbaa0f2758d6ab92117c7352db8ede170153c6bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f943aedb7ef6351aa6b2a46b3e84b6db2fae1145bd40f3b9aee6d897aa0d5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v79sj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:54Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.697289 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d40160e-075f-44b7-9d9d-3291e7774f27\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7079548fe9529583a1c129f57838c24cf06d1cb55a97c05c63a05d0cc2e7ec9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7762237003a5a256ff20a16a4869462e2e0dd1426aa391f7ec74ed9b1f572d4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1129 04:39:02.908489 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1129 04:39:02.908637 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 04:39:02.909677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703886220/tls.crt::/tmp/serving-cert-1703886220/tls.key\\\\\\\"\\\\nI1129 04:39:03.077876 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 04:39:03.080368 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 04:39:03.080390 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 04:39:03.080424 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 04:39:03.080432 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 04:39:03.086120 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 04:39:03.086240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086412 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 04:39:03.086502 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 04:39:03.086563 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 04:39:03.086622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1129 04:39:03.086141 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1129 04:39:03.086836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:38:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:54Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.714282 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:54Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.719107 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.719158 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.719169 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.719186 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.719196 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:54Z","lastTransitionTime":"2025-11-29T04:39:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.743485 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd007c61-4689-4179-95ad-44a9724e8e7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28499112c33332c23be39a11cdf7fc4e1946d04524a3f7c534087179a6620121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dac1c084955d874741d36709fa848c2c475a2d7108d4378b677bb13a88184105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72802532a47b381ead2507e1ae31c49852f9b1d332f3c08fb1e3a972a128eead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c81d94f098efec48c1ff8e172e476bf806ccec5777974eea9c0f3850e5fd40e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3d7f1ba1a408c54f0470fca40e307b70ba208dc7eeece599b684820e3f56c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d91c4f461a1912117cb8ed3a8087c5efaaf0a67455b721bfbeaad1c7dda8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5ce3c68997bc5859df929b844d823adf00352454bb278d57e90edcc44817cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5ce3c68997bc5859df929b844d823adf00352454bb278d57e90edcc44817cbe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T04:39:35Z\\\",\\\"message\\\":\\\"l\\\\u003e UUID: UUIDName:} {Op:mutate Table:NB_Global Row:map[] Rows:[] Columns:[] Mutations:[{Column:nb_cfg Mutator:+= Value:1}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {6011affd-30a6-4be6-872d-e4cf1ca780cf}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1129 04:39:35.599606 6479 services_controller.go:434] Service openshift-multus/multus-admission-controller retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{multus-admission-controller openshift-multus c9eea3b1-f918-4c62-9731-c809988317c1 4579 0 2025-02-23 05:21:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:multus-admission-controller] map[service.alpha.openshift.io/serving-cert-secret-name:multus-admission-controller-secret service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{operator.openshift.io/v1 Network cluster 8d01ddba-7e05-4639-926a-4485de3b6327 0xc007916917 0xc007916918}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:webhook,Protocol:TCP,Port:443,TargetPort:{0 6443 },NodePort:0,AppProtocol:nil,},ServicePort{Name:metrics,Protocol:TCP,Port:8443,TargetPort:{1 0 https},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: multus-admission-controller,},ClusterIP:10.217.5.119,Type:ClusterIP,ExternalIPs:[],Sessi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fht8l_openshift-ovn-kubernetes(bd007c61-4689-4179-95ad-44a9724e8e7e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d135e42b8886f1f118abf8f4ce4cf2875ac4dc6c148bfc0677cef6b3651562c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fht8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:54Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.754240 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82a40839-9eab-4511-8494-ee3d2c6ba0b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://186c216d29cdc56583cd5f39acc7cfd0558f60c668b1f5cbf039b566c1747e12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02c75c8691748a626cbef2de964cb1f4579e702276e8b02d805db0876a8fee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://624c4a09beeecd568d6045a720dfa28c15821d931021cedd8776da4bb36b557f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5b8064433d389b05e3e4744ef4c1f72512db615eae94e2a847bbd4cb2518f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d5b8064433d389b05e3e4744ef4c1f72512db615eae94e2a847bbd4cb2518f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:38:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:54Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.765601 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:54Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.777578 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7lb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55156232-4c80-45a7-aad5-a5fa2793e4ed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6e5473c80737a5b7b00e5874062e0485ed7a4d74632202090a9b12f331953b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d2dtw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7lb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:54Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.787584 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0847696a7b7926047dc62d75739b3b6a7a04cd58efd1c1c2c992d5ae8d20ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b2ec8f0d384066ba90eb42c1a0a925ade3722b2515b0592d6f2a8f5e533a0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-26t88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:54Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.798483 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-qx9wn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce8ce318-27e6-434d-96c8-591d250262fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c99mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c99mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-qx9wn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:54Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.816866 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22ed28a-3748-4eb9-b3a7-c02aa38f7e94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33c0540ccb097b15f65369ff537aef4ae621cc828ba314eb61f0a0037608cadf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c241d29297a73dce12bd0bc2863c56475eb35b8fe8c93991ba0ef8695e38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f42a55d797c6c94af3a5eac1bf08aa38a82359b7ddd0ba32dcc797218ceee6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebed9ec415bb433935da36df8014f9087236cea9788b2048cbff8714f8ad0df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:54Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.821360 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.821410 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.821422 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.821440 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.821453 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:54Z","lastTransitionTime":"2025-11-29T04:39:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.831834 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://deb9f370792598608cf2e0c0fe9975eaf8b95dc94b12eef8c55112df4aa9c998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:54Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.847167 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:54Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.859838 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee17b4a9a8999e5f76ceaf69dbd13acf80ed0e95e72f3e1428ce38c5d5982788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://382256e0fa1a4c26610227b9ae96288ccbfbbd97aa3cbd5fd4f18cb4228a234b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:54Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.874540 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt2dl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6c9236-5f7d-46c5-b080-4667620d874e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b6c827d0c57a87808fef301661fce28c50359e85c9e47947c50945f4f869de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt2dl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:54Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.889711 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s5d7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0af3d308-36b9-4622-8544-635f7b24d687\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23c8830e316099ef7b0e0044145a128658456871418c8c53bcc71210be290df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0088b60069e830e9c36075b09d1ff135da79e94fe6cd271d0e93415af50557e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0088b60069e830e9c36075b09d1ff135da79e94fe6cd271d0e93415af50557e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9b0b72fbe8b552d1266c310f6a6a67822c4c70dec1f812d31cc00748d50565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9b0b72fbe8b552d1266c310f6a6a67822c4c70dec1f812d31cc00748d50565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa3b85ae7c47440f0f61436e1cb1f3f2096e0e4011434876f9d17e913ca24df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa3b85ae7c47440f0f61436e1cb1f3f2096e0e4011434876f9d17e913ca24df7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s5d7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:54Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.901033 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d87rr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a92ae69-5675-49d9-a567-e2476fe02ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://207d152294f9996ed7294d858842da5eef891a6d8a027473dc3d29a7101aadab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffwm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d87rr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:54Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.923824 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.923867 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.923878 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.923892 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:54 crc kubenswrapper[4799]: I1129 04:39:54.923902 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:54Z","lastTransitionTime":"2025-11-29T04:39:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.026169 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.026205 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.026214 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.026228 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.026238 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:55Z","lastTransitionTime":"2025-11-29T04:39:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.129307 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.129357 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.129369 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.129387 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.129398 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:55Z","lastTransitionTime":"2025-11-29T04:39:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.231588 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.231633 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.231643 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.231657 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.231666 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:55Z","lastTransitionTime":"2025-11-29T04:39:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.334278 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.334350 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.334360 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.334377 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.334386 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:55Z","lastTransitionTime":"2025-11-29T04:39:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.436510 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.436594 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.436606 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.436630 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.436642 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:55Z","lastTransitionTime":"2025-11-29T04:39:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.539469 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.539516 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.539527 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.539543 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.539553 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:55Z","lastTransitionTime":"2025-11-29T04:39:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.643042 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.643118 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.643140 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.643173 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.643195 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:55Z","lastTransitionTime":"2025-11-29T04:39:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.658404 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.658428 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.658435 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qx9wn" Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.658582 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:39:55 crc kubenswrapper[4799]: E1129 04:39:55.658737 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 04:39:55 crc kubenswrapper[4799]: E1129 04:39:55.658879 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 04:39:55 crc kubenswrapper[4799]: E1129 04:39:55.659074 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qx9wn" podUID="ce8ce318-27e6-434d-96c8-591d250262fd" Nov 29 04:39:55 crc kubenswrapper[4799]: E1129 04:39:55.659119 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.746193 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.746245 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.746265 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.746283 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.746295 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:55Z","lastTransitionTime":"2025-11-29T04:39:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.848556 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.848594 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.848602 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.848618 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.848627 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:55Z","lastTransitionTime":"2025-11-29T04:39:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.908461 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ce8ce318-27e6-434d-96c8-591d250262fd-metrics-certs\") pod \"network-metrics-daemon-qx9wn\" (UID: \"ce8ce318-27e6-434d-96c8-591d250262fd\") " pod="openshift-multus/network-metrics-daemon-qx9wn" Nov 29 04:39:55 crc kubenswrapper[4799]: E1129 04:39:55.908648 4799 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 29 04:39:55 crc kubenswrapper[4799]: E1129 04:39:55.908733 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ce8ce318-27e6-434d-96c8-591d250262fd-metrics-certs podName:ce8ce318-27e6-434d-96c8-591d250262fd nodeName:}" failed. No retries permitted until 2025-11-29 04:40:27.908711159 +0000 UTC m=+103.551641759 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ce8ce318-27e6-434d-96c8-591d250262fd-metrics-certs") pod "network-metrics-daemon-qx9wn" (UID: "ce8ce318-27e6-434d-96c8-591d250262fd") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.950355 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.950386 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.950396 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.950410 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:55 crc kubenswrapper[4799]: I1129 04:39:55.950421 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:55Z","lastTransitionTime":"2025-11-29T04:39:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.052849 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.052893 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.052904 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.052921 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.052931 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:56Z","lastTransitionTime":"2025-11-29T04:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.155322 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.155357 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.155369 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.155383 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.155392 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:56Z","lastTransitionTime":"2025-11-29T04:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.208560 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-qt2dl_3b6c9236-5f7d-46c5-b080-4667620d874e/kube-multus/0.log" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.208612 4799 generic.go:334] "Generic (PLEG): container finished" podID="3b6c9236-5f7d-46c5-b080-4667620d874e" containerID="92b6c827d0c57a87808fef301661fce28c50359e85c9e47947c50945f4f869de" exitCode=1 Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.208643 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-qt2dl" event={"ID":"3b6c9236-5f7d-46c5-b080-4667620d874e","Type":"ContainerDied","Data":"92b6c827d0c57a87808fef301661fce28c50359e85c9e47947c50945f4f869de"} Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.209045 4799 scope.go:117] "RemoveContainer" containerID="92b6c827d0c57a87808fef301661fce28c50359e85c9e47947c50945f4f869de" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.223834 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-qx9wn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce8ce318-27e6-434d-96c8-591d250262fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c99mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c99mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-qx9wn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:56Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.239897 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22ed28a-3748-4eb9-b3a7-c02aa38f7e94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33c0540ccb097b15f65369ff537aef4ae621cc828ba314eb61f0a0037608cadf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c241d29297a73dce12bd0bc2863c56475eb35b8fe8c93991ba0ef8695e38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f42a55d797c6c94af3a5eac1bf08aa38a82359b7ddd0ba32dcc797218ceee6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebed9ec415bb433935da36df8014f9087236cea9788b2048cbff8714f8ad0df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:56Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.257064 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:56Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.258458 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.258492 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.258501 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.258517 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.258528 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:56Z","lastTransitionTime":"2025-11-29T04:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.271764 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7lb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55156232-4c80-45a7-aad5-a5fa2793e4ed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6e5473c80737a5b7b00e5874062e0485ed7a4d74632202090a9b12f331953b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d2dtw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7lb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:56Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.285228 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0847696a7b7926047dc62d75739b3b6a7a04cd58efd1c1c2c992d5ae8d20ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b2ec8f0d384066ba90eb42c1a0a925ade3722b2515b0592d6f2a8f5e533a0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-26t88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:56Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.306897 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s5d7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0af3d308-36b9-4622-8544-635f7b24d687\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23c8830e316099ef7b0e0044145a128658456871418c8c53bcc71210be290df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0088b60069e830e9c36075b09d1ff135da79e94fe6cd271d0e93415af50557e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0088b60069e830e9c36075b09d1ff135da79e94fe6cd271d0e93415af50557e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9b0b72fbe8b552d1266c310f6a6a67822c4c70dec1f812d31cc00748d50565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9b0b72fbe8b552d1266c310f6a6a67822c4c70dec1f812d31cc00748d50565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa3b85ae7c47440f0f61436e1cb1f3f2096e0e4011434876f9d17e913ca24df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa3b85ae7c47440f0f61436e1cb1f3f2096e0e4011434876f9d17e913ca24df7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s5d7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:56Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.317379 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d87rr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a92ae69-5675-49d9-a567-e2476fe02ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://207d152294f9996ed7294d858842da5eef891a6d8a027473dc3d29a7101aadab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffwm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d87rr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:56Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.334756 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://deb9f370792598608cf2e0c0fe9975eaf8b95dc94b12eef8c55112df4aa9c998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:56Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.347123 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:56Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.358753 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee17b4a9a8999e5f76ceaf69dbd13acf80ed0e95e72f3e1428ce38c5d5982788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://382256e0fa1a4c26610227b9ae96288ccbfbbd97aa3cbd5fd4f18cb4228a234b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:56Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.360809 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.360845 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.360857 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.360874 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.360885 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:56Z","lastTransitionTime":"2025-11-29T04:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.372871 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt2dl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6c9236-5f7d-46c5-b080-4667620d874e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b6c827d0c57a87808fef301661fce28c50359e85c9e47947c50945f4f869de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92b6c827d0c57a87808fef301661fce28c50359e85c9e47947c50945f4f869de\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T04:39:55Z\\\",\\\"message\\\":\\\"2025-11-29T04:39:10+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_797f994e-104f-4f0f-862f-201fc8b2144f\\\\n2025-11-29T04:39:10+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_797f994e-104f-4f0f-862f-201fc8b2144f to /host/opt/cni/bin/\\\\n2025-11-29T04:39:10Z [verbose] multus-daemon started\\\\n2025-11-29T04:39:10Z [verbose] Readiness Indicator file check\\\\n2025-11-29T04:39:55Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt2dl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:56Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.390907 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d40160e-075f-44b7-9d9d-3291e7774f27\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7079548fe9529583a1c129f57838c24cf06d1cb55a97c05c63a05d0cc2e7ec9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7762237003a5a256ff20a16a4869462e2e0dd1426aa391f7ec74ed9b1f572d4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1129 04:39:02.908489 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1129 04:39:02.908637 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 04:39:02.909677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703886220/tls.crt::/tmp/serving-cert-1703886220/tls.key\\\\\\\"\\\\nI1129 04:39:03.077876 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 04:39:03.080368 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 04:39:03.080390 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 04:39:03.080424 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 04:39:03.080432 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 04:39:03.086120 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 04:39:03.086240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086412 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 04:39:03.086502 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 04:39:03.086563 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 04:39:03.086622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1129 04:39:03.086141 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1129 04:39:03.086836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:38:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:56Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.404512 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://396c1c8cbd638c7dc4d8eebded6a17449d9abdc049161578581f1d8c3e31d9a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:56Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.422181 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v79sj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8cf268-e9fd-460a-b579-6411942a00a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c4e099a3149affd1f9f9c5cbaa0f2758d6ab92117c7352db8ede170153c6bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f943aedb7ef6351aa6b2a46b3e84b6db2fae1145bd40f3b9aee6d897aa0d5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v79sj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:56Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.440869 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82a40839-9eab-4511-8494-ee3d2c6ba0b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://186c216d29cdc56583cd5f39acc7cfd0558f60c668b1f5cbf039b566c1747e12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02c75c8691748a626cbef2de964cb1f4579e702276e8b02d805db0876a8fee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://624c4a09beeecd568d6045a720dfa28c15821d931021cedd8776da4bb36b557f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5b8064433d389b05e3e4744ef4c1f72512db615eae94e2a847bbd4cb2518f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d5b8064433d389b05e3e4744ef4c1f72512db615eae94e2a847bbd4cb2518f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:38:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:56Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.456296 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:56Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.463477 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.463534 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.463548 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.463571 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.463588 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:56Z","lastTransitionTime":"2025-11-29T04:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.487209 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd007c61-4689-4179-95ad-44a9724e8e7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28499112c33332c23be39a11cdf7fc4e1946d04524a3f7c534087179a6620121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dac1c084955d874741d36709fa848c2c475a2d7108d4378b677bb13a88184105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72802532a47b381ead2507e1ae31c49852f9b1d332f3c08fb1e3a972a128eead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c81d94f098efec48c1ff8e172e476bf806ccec5777974eea9c0f3850e5fd40e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3d7f1ba1a408c54f0470fca40e307b70ba208dc7eeece599b684820e3f56c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d91c4f461a1912117cb8ed3a8087c5efaaf0a67455b721bfbeaad1c7dda8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5ce3c68997bc5859df929b844d823adf00352454bb278d57e90edcc44817cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5ce3c68997bc5859df929b844d823adf00352454bb278d57e90edcc44817cbe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T04:39:35Z\\\",\\\"message\\\":\\\"l\\\\u003e UUID: UUIDName:} {Op:mutate Table:NB_Global Row:map[] Rows:[] Columns:[] Mutations:[{Column:nb_cfg Mutator:+= Value:1}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {6011affd-30a6-4be6-872d-e4cf1ca780cf}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1129 04:39:35.599606 6479 services_controller.go:434] Service openshift-multus/multus-admission-controller retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{multus-admission-controller openshift-multus c9eea3b1-f918-4c62-9731-c809988317c1 4579 0 2025-02-23 05:21:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:multus-admission-controller] map[service.alpha.openshift.io/serving-cert-secret-name:multus-admission-controller-secret service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{operator.openshift.io/v1 Network cluster 8d01ddba-7e05-4639-926a-4485de3b6327 0xc007916917 0xc007916918}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:webhook,Protocol:TCP,Port:443,TargetPort:{0 6443 },NodePort:0,AppProtocol:nil,},ServicePort{Name:metrics,Protocol:TCP,Port:8443,TargetPort:{1 0 https},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: multus-admission-controller,},ClusterIP:10.217.5.119,Type:ClusterIP,ExternalIPs:[],Sessi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fht8l_openshift-ovn-kubernetes(bd007c61-4689-4179-95ad-44a9724e8e7e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d135e42b8886f1f118abf8f4ce4cf2875ac4dc6c148bfc0677cef6b3651562c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fht8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:56Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.566480 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.566527 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.566537 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.566554 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.566565 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:56Z","lastTransitionTime":"2025-11-29T04:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.668896 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.668951 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.668964 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.668980 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.668992 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:56Z","lastTransitionTime":"2025-11-29T04:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.771646 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.771710 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.771728 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.771754 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.771772 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:56Z","lastTransitionTime":"2025-11-29T04:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.874770 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.874855 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.874873 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.874898 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.874915 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:56Z","lastTransitionTime":"2025-11-29T04:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.977878 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.977926 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.977935 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.977950 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:56 crc kubenswrapper[4799]: I1129 04:39:56.977978 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:56Z","lastTransitionTime":"2025-11-29T04:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.080509 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.080566 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.080585 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.080609 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.080625 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:57Z","lastTransitionTime":"2025-11-29T04:39:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.183367 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.183413 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.183426 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.183451 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.183464 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:57Z","lastTransitionTime":"2025-11-29T04:39:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.214752 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-qt2dl_3b6c9236-5f7d-46c5-b080-4667620d874e/kube-multus/0.log" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.214857 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-qt2dl" event={"ID":"3b6c9236-5f7d-46c5-b080-4667620d874e","Type":"ContainerStarted","Data":"d0e3c7d98a4a048822bc263c302eef3b130a53051289505324e512944f332b5d"} Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.228861 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22ed28a-3748-4eb9-b3a7-c02aa38f7e94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33c0540ccb097b15f65369ff537aef4ae621cc828ba314eb61f0a0037608cadf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c241d29297a73dce12bd0bc2863c56475eb35b8fe8c93991ba0ef8695e38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f42a55d797c6c94af3a5eac1bf08aa38a82359b7ddd0ba32dcc797218ceee6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebed9ec415bb433935da36df8014f9087236cea9788b2048cbff8714f8ad0df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:57Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.243659 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:57Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.254238 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7lb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55156232-4c80-45a7-aad5-a5fa2793e4ed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6e5473c80737a5b7b00e5874062e0485ed7a4d74632202090a9b12f331953b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d2dtw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7lb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:57Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.267265 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0847696a7b7926047dc62d75739b3b6a7a04cd58efd1c1c2c992d5ae8d20ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b2ec8f0d384066ba90eb42c1a0a925ade3722b2515b0592d6f2a8f5e533a0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-26t88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:57Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.282411 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-qx9wn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce8ce318-27e6-434d-96c8-591d250262fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c99mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c99mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-qx9wn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:57Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.286499 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.286544 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.286558 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.286583 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.286604 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:57Z","lastTransitionTime":"2025-11-29T04:39:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.295539 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://deb9f370792598608cf2e0c0fe9975eaf8b95dc94b12eef8c55112df4aa9c998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:57Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.307863 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:57Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.324514 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee17b4a9a8999e5f76ceaf69dbd13acf80ed0e95e72f3e1428ce38c5d5982788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://382256e0fa1a4c26610227b9ae96288ccbfbbd97aa3cbd5fd4f18cb4228a234b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:57Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.343954 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt2dl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6c9236-5f7d-46c5-b080-4667620d874e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0e3c7d98a4a048822bc263c302eef3b130a53051289505324e512944f332b5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92b6c827d0c57a87808fef301661fce28c50359e85c9e47947c50945f4f869de\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T04:39:55Z\\\",\\\"message\\\":\\\"2025-11-29T04:39:10+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_797f994e-104f-4f0f-862f-201fc8b2144f\\\\n2025-11-29T04:39:10+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_797f994e-104f-4f0f-862f-201fc8b2144f to /host/opt/cni/bin/\\\\n2025-11-29T04:39:10Z [verbose] multus-daemon started\\\\n2025-11-29T04:39:10Z [verbose] Readiness Indicator file check\\\\n2025-11-29T04:39:55Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt2dl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:57Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.364990 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s5d7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0af3d308-36b9-4622-8544-635f7b24d687\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23c8830e316099ef7b0e0044145a128658456871418c8c53bcc71210be290df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0088b60069e830e9c36075b09d1ff135da79e94fe6cd271d0e93415af50557e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0088b60069e830e9c36075b09d1ff135da79e94fe6cd271d0e93415af50557e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9b0b72fbe8b552d1266c310f6a6a67822c4c70dec1f812d31cc00748d50565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9b0b72fbe8b552d1266c310f6a6a67822c4c70dec1f812d31cc00748d50565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa3b85ae7c47440f0f61436e1cb1f3f2096e0e4011434876f9d17e913ca24df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa3b85ae7c47440f0f61436e1cb1f3f2096e0e4011434876f9d17e913ca24df7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s5d7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:57Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.376969 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d87rr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a92ae69-5675-49d9-a567-e2476fe02ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://207d152294f9996ed7294d858842da5eef891a6d8a027473dc3d29a7101aadab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffwm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d87rr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:57Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.389561 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.389615 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.389626 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.389642 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.389653 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:57Z","lastTransitionTime":"2025-11-29T04:39:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.397373 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d40160e-075f-44b7-9d9d-3291e7774f27\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7079548fe9529583a1c129f57838c24cf06d1cb55a97c05c63a05d0cc2e7ec9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7762237003a5a256ff20a16a4869462e2e0dd1426aa391f7ec74ed9b1f572d4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1129 04:39:02.908489 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1129 04:39:02.908637 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 04:39:02.909677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703886220/tls.crt::/tmp/serving-cert-1703886220/tls.key\\\\\\\"\\\\nI1129 04:39:03.077876 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 04:39:03.080368 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 04:39:03.080390 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 04:39:03.080424 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 04:39:03.080432 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 04:39:03.086120 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 04:39:03.086240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086412 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 04:39:03.086502 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 04:39:03.086563 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 04:39:03.086622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1129 04:39:03.086141 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1129 04:39:03.086836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:38:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:57Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.411480 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://396c1c8cbd638c7dc4d8eebded6a17449d9abdc049161578581f1d8c3e31d9a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:57Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.422281 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v79sj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8cf268-e9fd-460a-b579-6411942a00a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c4e099a3149affd1f9f9c5cbaa0f2758d6ab92117c7352db8ede170153c6bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f943aedb7ef6351aa6b2a46b3e84b6db2fae1145bd40f3b9aee6d897aa0d5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v79sj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:57Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.438432 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82a40839-9eab-4511-8494-ee3d2c6ba0b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://186c216d29cdc56583cd5f39acc7cfd0558f60c668b1f5cbf039b566c1747e12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02c75c8691748a626cbef2de964cb1f4579e702276e8b02d805db0876a8fee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://624c4a09beeecd568d6045a720dfa28c15821d931021cedd8776da4bb36b557f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5b8064433d389b05e3e4744ef4c1f72512db615eae94e2a847bbd4cb2518f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d5b8064433d389b05e3e4744ef4c1f72512db615eae94e2a847bbd4cb2518f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:38:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:57Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.450903 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:57Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.468654 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd007c61-4689-4179-95ad-44a9724e8e7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28499112c33332c23be39a11cdf7fc4e1946d04524a3f7c534087179a6620121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dac1c084955d874741d36709fa848c2c475a2d7108d4378b677bb13a88184105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72802532a47b381ead2507e1ae31c49852f9b1d332f3c08fb1e3a972a128eead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c81d94f098efec48c1ff8e172e476bf806ccec5777974eea9c0f3850e5fd40e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3d7f1ba1a408c54f0470fca40e307b70ba208dc7eeece599b684820e3f56c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d91c4f461a1912117cb8ed3a8087c5efaaf0a67455b721bfbeaad1c7dda8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5ce3c68997bc5859df929b844d823adf00352454bb278d57e90edcc44817cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5ce3c68997bc5859df929b844d823adf00352454bb278d57e90edcc44817cbe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T04:39:35Z\\\",\\\"message\\\":\\\"l\\\\u003e UUID: UUIDName:} {Op:mutate Table:NB_Global Row:map[] Rows:[] Columns:[] Mutations:[{Column:nb_cfg Mutator:+= Value:1}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {6011affd-30a6-4be6-872d-e4cf1ca780cf}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1129 04:39:35.599606 6479 services_controller.go:434] Service openshift-multus/multus-admission-controller retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{multus-admission-controller openshift-multus c9eea3b1-f918-4c62-9731-c809988317c1 4579 0 2025-02-23 05:21:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:multus-admission-controller] map[service.alpha.openshift.io/serving-cert-secret-name:multus-admission-controller-secret service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{operator.openshift.io/v1 Network cluster 8d01ddba-7e05-4639-926a-4485de3b6327 0xc007916917 0xc007916918}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:webhook,Protocol:TCP,Port:443,TargetPort:{0 6443 },NodePort:0,AppProtocol:nil,},ServicePort{Name:metrics,Protocol:TCP,Port:8443,TargetPort:{1 0 https},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: multus-admission-controller,},ClusterIP:10.217.5.119,Type:ClusterIP,ExternalIPs:[],Sessi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fht8l_openshift-ovn-kubernetes(bd007c61-4689-4179-95ad-44a9724e8e7e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d135e42b8886f1f118abf8f4ce4cf2875ac4dc6c148bfc0677cef6b3651562c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fht8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:57Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.498848 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.498951 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.498975 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.499007 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.499026 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:57Z","lastTransitionTime":"2025-11-29T04:39:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.602862 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.602930 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.602949 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.602979 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.603002 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:57Z","lastTransitionTime":"2025-11-29T04:39:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.658186 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qx9wn" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.658281 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.658190 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.658185 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:39:57 crc kubenswrapper[4799]: E1129 04:39:57.658413 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qx9wn" podUID="ce8ce318-27e6-434d-96c8-591d250262fd" Nov 29 04:39:57 crc kubenswrapper[4799]: E1129 04:39:57.658674 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 04:39:57 crc kubenswrapper[4799]: E1129 04:39:57.658964 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 04:39:57 crc kubenswrapper[4799]: E1129 04:39:57.659226 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.706113 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.706151 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.706161 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.706176 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.706187 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:57Z","lastTransitionTime":"2025-11-29T04:39:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.808834 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.808877 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.808887 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.808902 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.808913 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:57Z","lastTransitionTime":"2025-11-29T04:39:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.911354 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.911396 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.911406 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.911421 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:57 crc kubenswrapper[4799]: I1129 04:39:57.911432 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:57Z","lastTransitionTime":"2025-11-29T04:39:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.013757 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.013812 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.013824 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.013839 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.013851 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:58Z","lastTransitionTime":"2025-11-29T04:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.115675 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.115735 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.115756 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.115781 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.115830 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:58Z","lastTransitionTime":"2025-11-29T04:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.217699 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.217747 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.217759 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.217774 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.217813 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:58Z","lastTransitionTime":"2025-11-29T04:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.320862 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.320901 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.320912 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.320928 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.320941 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:58Z","lastTransitionTime":"2025-11-29T04:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.423694 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.423743 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.423754 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.423771 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.423783 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:58Z","lastTransitionTime":"2025-11-29T04:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.526614 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.526691 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.526712 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.526734 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.526751 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:58Z","lastTransitionTime":"2025-11-29T04:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.630041 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.630158 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.630180 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.630206 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.630223 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:58Z","lastTransitionTime":"2025-11-29T04:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.732856 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.732932 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.732957 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.732987 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.733011 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:58Z","lastTransitionTime":"2025-11-29T04:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.806871 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.806938 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.806963 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.806990 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.807006 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:58Z","lastTransitionTime":"2025-11-29T04:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:58 crc kubenswrapper[4799]: E1129 04:39:58.823765 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6f4e4672-da32-440c-8c4f-6e7f4942f55e\\\",\\\"systemUUID\\\":\\\"882c4d8e-57b7-4312-8499-40d1f85d56c9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:58Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.829178 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.829238 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.829260 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.829293 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.829335 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:58Z","lastTransitionTime":"2025-11-29T04:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:58 crc kubenswrapper[4799]: E1129 04:39:58.852109 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6f4e4672-da32-440c-8c4f-6e7f4942f55e\\\",\\\"systemUUID\\\":\\\"882c4d8e-57b7-4312-8499-40d1f85d56c9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:58Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.857891 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.857969 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.857992 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.858024 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.858049 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:58Z","lastTransitionTime":"2025-11-29T04:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:58 crc kubenswrapper[4799]: E1129 04:39:58.881638 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6f4e4672-da32-440c-8c4f-6e7f4942f55e\\\",\\\"systemUUID\\\":\\\"882c4d8e-57b7-4312-8499-40d1f85d56c9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:58Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.888023 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.888138 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.888169 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.888202 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.888229 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:58Z","lastTransitionTime":"2025-11-29T04:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:58 crc kubenswrapper[4799]: E1129 04:39:58.908228 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6f4e4672-da32-440c-8c4f-6e7f4942f55e\\\",\\\"systemUUID\\\":\\\"882c4d8e-57b7-4312-8499-40d1f85d56c9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:58Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.919383 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.919590 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.919629 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.919661 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.919685 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:58Z","lastTransitionTime":"2025-11-29T04:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:58 crc kubenswrapper[4799]: E1129 04:39:58.940647 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T04:39:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6f4e4672-da32-440c-8c4f-6e7f4942f55e\\\",\\\"systemUUID\\\":\\\"882c4d8e-57b7-4312-8499-40d1f85d56c9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:39:58Z is after 2025-08-24T17:21:41Z" Nov 29 04:39:58 crc kubenswrapper[4799]: E1129 04:39:58.940817 4799 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.942942 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.942996 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.943029 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.943050 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:58 crc kubenswrapper[4799]: I1129 04:39:58.943066 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:58Z","lastTransitionTime":"2025-11-29T04:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.046601 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.046646 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.046655 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.046674 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.046686 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:59Z","lastTransitionTime":"2025-11-29T04:39:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.149613 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.149681 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.149700 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.149727 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.149747 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:59Z","lastTransitionTime":"2025-11-29T04:39:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.254114 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.254230 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.254287 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.254317 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.254376 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:59Z","lastTransitionTime":"2025-11-29T04:39:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.359456 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.359538 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.359565 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.359603 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.359625 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:59Z","lastTransitionTime":"2025-11-29T04:39:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.463704 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.463774 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.463830 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.463870 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.463891 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:59Z","lastTransitionTime":"2025-11-29T04:39:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.567456 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.567539 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.567566 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.567601 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.567625 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:59Z","lastTransitionTime":"2025-11-29T04:39:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.658245 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.658433 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qx9wn" Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.658245 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 04:39:59 crc kubenswrapper[4799]: E1129 04:39:59.658547 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.658267 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 04:39:59 crc kubenswrapper[4799]: E1129 04:39:59.658820 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qx9wn" podUID="ce8ce318-27e6-434d-96c8-591d250262fd" Nov 29 04:39:59 crc kubenswrapper[4799]: E1129 04:39:59.658911 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 04:39:59 crc kubenswrapper[4799]: E1129 04:39:59.659046 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.670712 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.670763 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.670775 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.670813 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.670830 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:59Z","lastTransitionTime":"2025-11-29T04:39:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.775493 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.775572 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.775600 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.775630 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.775652 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:59Z","lastTransitionTime":"2025-11-29T04:39:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.879159 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.879233 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.879254 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.879286 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.879307 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:59Z","lastTransitionTime":"2025-11-29T04:39:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.982680 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.982756 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.982775 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.982829 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:39:59 crc kubenswrapper[4799]: I1129 04:39:59.982848 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:39:59Z","lastTransitionTime":"2025-11-29T04:39:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:00 crc kubenswrapper[4799]: I1129 04:40:00.086101 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:00 crc kubenswrapper[4799]: I1129 04:40:00.086156 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:00 crc kubenswrapper[4799]: I1129 04:40:00.086176 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:00 crc kubenswrapper[4799]: I1129 04:40:00.086203 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:00 crc kubenswrapper[4799]: I1129 04:40:00.086221 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:00Z","lastTransitionTime":"2025-11-29T04:40:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:00 crc kubenswrapper[4799]: I1129 04:40:00.189541 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:00 crc kubenswrapper[4799]: I1129 04:40:00.189596 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:00 crc kubenswrapper[4799]: I1129 04:40:00.189610 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:00 crc kubenswrapper[4799]: I1129 04:40:00.189627 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:00 crc kubenswrapper[4799]: I1129 04:40:00.189639 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:00Z","lastTransitionTime":"2025-11-29T04:40:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:00 crc kubenswrapper[4799]: I1129 04:40:00.292279 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:00 crc kubenswrapper[4799]: I1129 04:40:00.292372 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:00 crc kubenswrapper[4799]: I1129 04:40:00.292398 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:00 crc kubenswrapper[4799]: I1129 04:40:00.292436 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:00 crc kubenswrapper[4799]: I1129 04:40:00.292463 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:00Z","lastTransitionTime":"2025-11-29T04:40:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:00 crc kubenswrapper[4799]: I1129 04:40:00.395220 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:00 crc kubenswrapper[4799]: I1129 04:40:00.395296 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:00 crc kubenswrapper[4799]: I1129 04:40:00.395313 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:00 crc kubenswrapper[4799]: I1129 04:40:00.395357 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:00 crc kubenswrapper[4799]: I1129 04:40:00.395370 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:00Z","lastTransitionTime":"2025-11-29T04:40:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:00 crc kubenswrapper[4799]: I1129 04:40:00.499416 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:00 crc kubenswrapper[4799]: I1129 04:40:00.499518 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:00 crc kubenswrapper[4799]: I1129 04:40:00.499544 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:00 crc kubenswrapper[4799]: I1129 04:40:00.499585 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:00 crc kubenswrapper[4799]: I1129 04:40:00.499614 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:00Z","lastTransitionTime":"2025-11-29T04:40:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:00 crc kubenswrapper[4799]: I1129 04:40:00.602474 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:00 crc kubenswrapper[4799]: I1129 04:40:00.602521 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:00 crc kubenswrapper[4799]: I1129 04:40:00.602535 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:00 crc kubenswrapper[4799]: I1129 04:40:00.602552 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:00 crc kubenswrapper[4799]: I1129 04:40:00.602564 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:00Z","lastTransitionTime":"2025-11-29T04:40:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:00 crc kubenswrapper[4799]: I1129 04:40:00.704909 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:00 crc kubenswrapper[4799]: I1129 04:40:00.704969 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:00 crc kubenswrapper[4799]: I1129 04:40:00.704986 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:00 crc kubenswrapper[4799]: I1129 04:40:00.705006 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:00 crc kubenswrapper[4799]: I1129 04:40:00.705020 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:00Z","lastTransitionTime":"2025-11-29T04:40:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:00 crc kubenswrapper[4799]: I1129 04:40:00.808488 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:00 crc kubenswrapper[4799]: I1129 04:40:00.808561 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:00 crc kubenswrapper[4799]: I1129 04:40:00.808583 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:00 crc kubenswrapper[4799]: I1129 04:40:00.808609 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:00 crc kubenswrapper[4799]: I1129 04:40:00.808635 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:00Z","lastTransitionTime":"2025-11-29T04:40:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:00 crc kubenswrapper[4799]: I1129 04:40:00.911614 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:00 crc kubenswrapper[4799]: I1129 04:40:00.911977 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:00 crc kubenswrapper[4799]: I1129 04:40:00.912173 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:00 crc kubenswrapper[4799]: I1129 04:40:00.912309 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:00 crc kubenswrapper[4799]: I1129 04:40:00.912468 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:00Z","lastTransitionTime":"2025-11-29T04:40:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.015389 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.015886 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.016059 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.016213 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.016420 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:01Z","lastTransitionTime":"2025-11-29T04:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.120545 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.120607 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.120625 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.120676 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.120694 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:01Z","lastTransitionTime":"2025-11-29T04:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.230467 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.230501 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.230511 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.230524 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.230533 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:01Z","lastTransitionTime":"2025-11-29T04:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.333243 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.333276 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.333284 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.333297 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.333306 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:01Z","lastTransitionTime":"2025-11-29T04:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.436469 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.437161 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.437963 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.438240 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.438417 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:01Z","lastTransitionTime":"2025-11-29T04:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.541414 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.541901 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.542088 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.542249 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.542434 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:01Z","lastTransitionTime":"2025-11-29T04:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.645780 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.645896 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.645921 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.645948 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.645965 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:01Z","lastTransitionTime":"2025-11-29T04:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.658494 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.658554 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qx9wn" Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.658616 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:40:01 crc kubenswrapper[4799]: E1129 04:40:01.658643 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.658496 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 04:40:01 crc kubenswrapper[4799]: E1129 04:40:01.658757 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 04:40:01 crc kubenswrapper[4799]: E1129 04:40:01.659012 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qx9wn" podUID="ce8ce318-27e6-434d-96c8-591d250262fd" Nov 29 04:40:01 crc kubenswrapper[4799]: E1129 04:40:01.659153 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.749577 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.749637 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.749660 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.749689 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.749711 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:01Z","lastTransitionTime":"2025-11-29T04:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.853587 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.853946 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.854235 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.854478 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.854739 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:01Z","lastTransitionTime":"2025-11-29T04:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.957642 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.958097 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.958176 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.958273 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:01 crc kubenswrapper[4799]: I1129 04:40:01.958354 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:01Z","lastTransitionTime":"2025-11-29T04:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:02 crc kubenswrapper[4799]: I1129 04:40:02.062246 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:02 crc kubenswrapper[4799]: I1129 04:40:02.062319 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:02 crc kubenswrapper[4799]: I1129 04:40:02.062339 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:02 crc kubenswrapper[4799]: I1129 04:40:02.062370 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:02 crc kubenswrapper[4799]: I1129 04:40:02.062393 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:02Z","lastTransitionTime":"2025-11-29T04:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:02 crc kubenswrapper[4799]: I1129 04:40:02.164832 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:02 crc kubenswrapper[4799]: I1129 04:40:02.164912 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:02 crc kubenswrapper[4799]: I1129 04:40:02.164931 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:02 crc kubenswrapper[4799]: I1129 04:40:02.164957 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:02 crc kubenswrapper[4799]: I1129 04:40:02.164969 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:02Z","lastTransitionTime":"2025-11-29T04:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:02 crc kubenswrapper[4799]: I1129 04:40:02.268583 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:02 crc kubenswrapper[4799]: I1129 04:40:02.268645 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:02 crc kubenswrapper[4799]: I1129 04:40:02.268657 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:02 crc kubenswrapper[4799]: I1129 04:40:02.268682 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:02 crc kubenswrapper[4799]: I1129 04:40:02.268696 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:02Z","lastTransitionTime":"2025-11-29T04:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:02 crc kubenswrapper[4799]: I1129 04:40:02.371740 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:02 crc kubenswrapper[4799]: I1129 04:40:02.371853 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:02 crc kubenswrapper[4799]: I1129 04:40:02.371901 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:02 crc kubenswrapper[4799]: I1129 04:40:02.371923 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:02 crc kubenswrapper[4799]: I1129 04:40:02.371939 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:02Z","lastTransitionTime":"2025-11-29T04:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:02 crc kubenswrapper[4799]: I1129 04:40:02.474295 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:02 crc kubenswrapper[4799]: I1129 04:40:02.474372 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:02 crc kubenswrapper[4799]: I1129 04:40:02.474388 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:02 crc kubenswrapper[4799]: I1129 04:40:02.474408 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:02 crc kubenswrapper[4799]: I1129 04:40:02.474423 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:02Z","lastTransitionTime":"2025-11-29T04:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:02 crc kubenswrapper[4799]: I1129 04:40:02.576926 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:02 crc kubenswrapper[4799]: I1129 04:40:02.576968 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:02 crc kubenswrapper[4799]: I1129 04:40:02.576978 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:02 crc kubenswrapper[4799]: I1129 04:40:02.576990 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:02 crc kubenswrapper[4799]: I1129 04:40:02.576998 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:02Z","lastTransitionTime":"2025-11-29T04:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:02 crc kubenswrapper[4799]: I1129 04:40:02.679472 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:02 crc kubenswrapper[4799]: I1129 04:40:02.679510 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:02 crc kubenswrapper[4799]: I1129 04:40:02.679520 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:02 crc kubenswrapper[4799]: I1129 04:40:02.679532 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:02 crc kubenswrapper[4799]: I1129 04:40:02.679541 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:02Z","lastTransitionTime":"2025-11-29T04:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:02 crc kubenswrapper[4799]: I1129 04:40:02.781955 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:02 crc kubenswrapper[4799]: I1129 04:40:02.782038 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:02 crc kubenswrapper[4799]: I1129 04:40:02.782062 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:02 crc kubenswrapper[4799]: I1129 04:40:02.782100 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:02 crc kubenswrapper[4799]: I1129 04:40:02.782122 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:02Z","lastTransitionTime":"2025-11-29T04:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:02 crc kubenswrapper[4799]: I1129 04:40:02.884861 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:02 crc kubenswrapper[4799]: I1129 04:40:02.884908 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:02 crc kubenswrapper[4799]: I1129 04:40:02.884925 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:02 crc kubenswrapper[4799]: I1129 04:40:02.885004 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:02 crc kubenswrapper[4799]: I1129 04:40:02.885056 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:02Z","lastTransitionTime":"2025-11-29T04:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:02 crc kubenswrapper[4799]: I1129 04:40:02.988685 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:02 crc kubenswrapper[4799]: I1129 04:40:02.988753 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:02 crc kubenswrapper[4799]: I1129 04:40:02.988772 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:02 crc kubenswrapper[4799]: I1129 04:40:02.988827 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:02 crc kubenswrapper[4799]: I1129 04:40:02.988847 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:02Z","lastTransitionTime":"2025-11-29T04:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:03 crc kubenswrapper[4799]: I1129 04:40:03.091387 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:03 crc kubenswrapper[4799]: I1129 04:40:03.091458 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:03 crc kubenswrapper[4799]: I1129 04:40:03.091489 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:03 crc kubenswrapper[4799]: I1129 04:40:03.091518 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:03 crc kubenswrapper[4799]: I1129 04:40:03.091536 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:03Z","lastTransitionTime":"2025-11-29T04:40:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:03 crc kubenswrapper[4799]: I1129 04:40:03.195013 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:03 crc kubenswrapper[4799]: I1129 04:40:03.195056 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:03 crc kubenswrapper[4799]: I1129 04:40:03.195065 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:03 crc kubenswrapper[4799]: I1129 04:40:03.195078 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:03 crc kubenswrapper[4799]: I1129 04:40:03.195092 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:03Z","lastTransitionTime":"2025-11-29T04:40:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:03 crc kubenswrapper[4799]: I1129 04:40:03.298383 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:03 crc kubenswrapper[4799]: I1129 04:40:03.298416 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:03 crc kubenswrapper[4799]: I1129 04:40:03.298425 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:03 crc kubenswrapper[4799]: I1129 04:40:03.298438 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:03 crc kubenswrapper[4799]: I1129 04:40:03.298446 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:03Z","lastTransitionTime":"2025-11-29T04:40:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:03 crc kubenswrapper[4799]: I1129 04:40:03.401594 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:03 crc kubenswrapper[4799]: I1129 04:40:03.401656 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:03 crc kubenswrapper[4799]: I1129 04:40:03.401669 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:03 crc kubenswrapper[4799]: I1129 04:40:03.401687 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:03 crc kubenswrapper[4799]: I1129 04:40:03.401699 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:03Z","lastTransitionTime":"2025-11-29T04:40:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:03 crc kubenswrapper[4799]: I1129 04:40:03.504356 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:03 crc kubenswrapper[4799]: I1129 04:40:03.504411 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:03 crc kubenswrapper[4799]: I1129 04:40:03.504428 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:03 crc kubenswrapper[4799]: I1129 04:40:03.504449 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:03 crc kubenswrapper[4799]: I1129 04:40:03.504464 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:03Z","lastTransitionTime":"2025-11-29T04:40:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:03 crc kubenswrapper[4799]: I1129 04:40:03.608291 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:03 crc kubenswrapper[4799]: I1129 04:40:03.608333 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:03 crc kubenswrapper[4799]: I1129 04:40:03.608351 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:03 crc kubenswrapper[4799]: I1129 04:40:03.608372 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:03 crc kubenswrapper[4799]: I1129 04:40:03.608389 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:03Z","lastTransitionTime":"2025-11-29T04:40:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:03 crc kubenswrapper[4799]: I1129 04:40:03.658589 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qx9wn" Nov 29 04:40:03 crc kubenswrapper[4799]: I1129 04:40:03.658589 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 04:40:03 crc kubenswrapper[4799]: E1129 04:40:03.658741 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qx9wn" podUID="ce8ce318-27e6-434d-96c8-591d250262fd" Nov 29 04:40:03 crc kubenswrapper[4799]: I1129 04:40:03.658618 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 04:40:03 crc kubenswrapper[4799]: I1129 04:40:03.658610 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:40:03 crc kubenswrapper[4799]: E1129 04:40:03.658854 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 04:40:03 crc kubenswrapper[4799]: E1129 04:40:03.658910 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 04:40:03 crc kubenswrapper[4799]: E1129 04:40:03.658962 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 04:40:03 crc kubenswrapper[4799]: I1129 04:40:03.710456 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:03 crc kubenswrapper[4799]: I1129 04:40:03.710489 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:03 crc kubenswrapper[4799]: I1129 04:40:03.710497 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:03 crc kubenswrapper[4799]: I1129 04:40:03.710511 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:03 crc kubenswrapper[4799]: I1129 04:40:03.710521 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:03Z","lastTransitionTime":"2025-11-29T04:40:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:03 crc kubenswrapper[4799]: I1129 04:40:03.812822 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:03 crc kubenswrapper[4799]: I1129 04:40:03.812861 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:03 crc kubenswrapper[4799]: I1129 04:40:03.812874 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:03 crc kubenswrapper[4799]: I1129 04:40:03.812893 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:03 crc kubenswrapper[4799]: I1129 04:40:03.812903 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:03Z","lastTransitionTime":"2025-11-29T04:40:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:03 crc kubenswrapper[4799]: I1129 04:40:03.915433 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:03 crc kubenswrapper[4799]: I1129 04:40:03.915507 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:03 crc kubenswrapper[4799]: I1129 04:40:03.915529 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:03 crc kubenswrapper[4799]: I1129 04:40:03.915555 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:03 crc kubenswrapper[4799]: I1129 04:40:03.915577 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:03Z","lastTransitionTime":"2025-11-29T04:40:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.018279 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.018320 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.018332 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.018347 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.018359 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:04Z","lastTransitionTime":"2025-11-29T04:40:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.120442 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.120480 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.120491 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.120506 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.120515 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:04Z","lastTransitionTime":"2025-11-29T04:40:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.222366 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.222413 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.222425 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.222444 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.222456 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:04Z","lastTransitionTime":"2025-11-29T04:40:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.325082 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.325131 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.325146 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.325167 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.325182 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:04Z","lastTransitionTime":"2025-11-29T04:40:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.427610 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.427707 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.427720 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.427736 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.427745 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:04Z","lastTransitionTime":"2025-11-29T04:40:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.530501 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.530561 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.530574 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.530592 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.530609 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:04Z","lastTransitionTime":"2025-11-29T04:40:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.633723 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.633856 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.633882 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.633913 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.633936 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:04Z","lastTransitionTime":"2025-11-29T04:40:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.682059 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://deb9f370792598608cf2e0c0fe9975eaf8b95dc94b12eef8c55112df4aa9c998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:40:04Z is after 2025-08-24T17:21:41Z" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.701844 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:40:04Z is after 2025-08-24T17:21:41Z" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.722088 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee17b4a9a8999e5f76ceaf69dbd13acf80ed0e95e72f3e1428ce38c5d5982788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://382256e0fa1a4c26610227b9ae96288ccbfbbd97aa3cbd5fd4f18cb4228a234b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:40:04Z is after 2025-08-24T17:21:41Z" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.737677 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.737759 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.737779 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.737836 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.737856 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:04Z","lastTransitionTime":"2025-11-29T04:40:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.742023 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt2dl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6c9236-5f7d-46c5-b080-4667620d874e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0e3c7d98a4a048822bc263c302eef3b130a53051289505324e512944f332b5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92b6c827d0c57a87808fef301661fce28c50359e85c9e47947c50945f4f869de\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T04:39:55Z\\\",\\\"message\\\":\\\"2025-11-29T04:39:10+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_797f994e-104f-4f0f-862f-201fc8b2144f\\\\n2025-11-29T04:39:10+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_797f994e-104f-4f0f-862f-201fc8b2144f to /host/opt/cni/bin/\\\\n2025-11-29T04:39:10Z [verbose] multus-daemon started\\\\n2025-11-29T04:39:10Z [verbose] Readiness Indicator file check\\\\n2025-11-29T04:39:55Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt2dl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:40:04Z is after 2025-08-24T17:21:41Z" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.769343 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s5d7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0af3d308-36b9-4622-8544-635f7b24d687\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23c8830e316099ef7b0e0044145a128658456871418c8c53bcc71210be290df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75d3bbc421996215d78626d1c7f59a9aafd1595c7fd301fba4db79630f157567\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc85654dbba23617adbe92ff595ddcb4f54128d9bf0837efadf7bdebb5d80a0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3558d878c31fb804a374503601a5ce6b7d39cc1c1ba24d246fb98f632832df46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0088b60069e830e9c36075b09d1ff135da79e94fe6cd271d0e93415af50557e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0088b60069e830e9c36075b09d1ff135da79e94fe6cd271d0e93415af50557e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9b0b72fbe8b552d1266c310f6a6a67822c4c70dec1f812d31cc00748d50565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9b0b72fbe8b552d1266c310f6a6a67822c4c70dec1f812d31cc00748d50565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa3b85ae7c47440f0f61436e1cb1f3f2096e0e4011434876f9d17e913ca24df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa3b85ae7c47440f0f61436e1cb1f3f2096e0e4011434876f9d17e913ca24df7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlq24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s5d7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:40:04Z is after 2025-08-24T17:21:41Z" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.789718 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d87rr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a92ae69-5675-49d9-a567-e2476fe02ef2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://207d152294f9996ed7294d858842da5eef891a6d8a027473dc3d29a7101aadab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffwm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d87rr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:40:04Z is after 2025-08-24T17:21:41Z" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.807595 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://396c1c8cbd638c7dc4d8eebded6a17449d9abdc049161578581f1d8c3e31d9a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:40:04Z is after 2025-08-24T17:21:41Z" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.826052 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v79sj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8cf268-e9fd-460a-b579-6411942a00a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c4e099a3149affd1f9f9c5cbaa0f2758d6ab92117c7352db8ede170153c6bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f943aedb7ef6351aa6b2a46b3e84b6db2fae1145bd40f3b9aee6d897aa0d5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v79sj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:40:04Z is after 2025-08-24T17:21:41Z" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.839903 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.839965 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.839983 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.840006 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.840027 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:04Z","lastTransitionTime":"2025-11-29T04:40:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.841487 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d40160e-075f-44b7-9d9d-3291e7774f27\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7079548fe9529583a1c129f57838c24cf06d1cb55a97c05c63a05d0cc2e7ec9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7762237003a5a256ff20a16a4869462e2e0dd1426aa391f7ec74ed9b1f572d4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1129 04:39:02.908489 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1129 04:39:02.908637 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 04:39:02.909677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703886220/tls.crt::/tmp/serving-cert-1703886220/tls.key\\\\\\\"\\\\nI1129 04:39:03.077876 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 04:39:03.080368 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 04:39:03.080390 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 04:39:03.080424 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 04:39:03.080432 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 04:39:03.086120 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 04:39:03.086240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 04:39:03.086412 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 04:39:03.086502 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 04:39:03.086563 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 04:39:03.086622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1129 04:39:03.086141 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1129 04:39:03.086836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:38:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:40:04Z is after 2025-08-24T17:21:41Z" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.858651 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:40:04Z is after 2025-08-24T17:21:41Z" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.888954 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd007c61-4689-4179-95ad-44a9724e8e7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28499112c33332c23be39a11cdf7fc4e1946d04524a3f7c534087179a6620121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dac1c084955d874741d36709fa848c2c475a2d7108d4378b677bb13a88184105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72802532a47b381ead2507e1ae31c49852f9b1d332f3c08fb1e3a972a128eead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c81d94f098efec48c1ff8e172e476bf806ccec5777974eea9c0f3850e5fd40e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3d7f1ba1a408c54f0470fca40e307b70ba208dc7eeece599b684820e3f56c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e76d91c4f461a1912117cb8ed3a8087c5efaaf0a67455b721bfbeaad1c7dda8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5ce3c68997bc5859df929b844d823adf00352454bb278d57e90edcc44817cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5ce3c68997bc5859df929b844d823adf00352454bb278d57e90edcc44817cbe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T04:39:35Z\\\",\\\"message\\\":\\\"l\\\\u003e UUID: UUIDName:} {Op:mutate Table:NB_Global Row:map[] Rows:[] Columns:[] Mutations:[{Column:nb_cfg Mutator:+= Value:1}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {6011affd-30a6-4be6-872d-e4cf1ca780cf}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1129 04:39:35.599606 6479 services_controller.go:434] Service openshift-multus/multus-admission-controller retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{multus-admission-controller openshift-multus c9eea3b1-f918-4c62-9731-c809988317c1 4579 0 2025-02-23 05:21:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:multus-admission-controller] map[service.alpha.openshift.io/serving-cert-secret-name:multus-admission-controller-secret service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{operator.openshift.io/v1 Network cluster 8d01ddba-7e05-4639-926a-4485de3b6327 0xc007916917 0xc007916918}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:webhook,Protocol:TCP,Port:443,TargetPort:{0 6443 },NodePort:0,AppProtocol:nil,},ServicePort{Name:metrics,Protocol:TCP,Port:8443,TargetPort:{1 0 https},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: multus-admission-controller,},ClusterIP:10.217.5.119,Type:ClusterIP,ExternalIPs:[],Sessi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fht8l_openshift-ovn-kubernetes(bd007c61-4689-4179-95ad-44a9724e8e7e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d135e42b8886f1f118abf8f4ce4cf2875ac4dc6c148bfc0677cef6b3651562c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jpd4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fht8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:40:04Z is after 2025-08-24T17:21:41Z" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.899842 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82a40839-9eab-4511-8494-ee3d2c6ba0b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://186c216d29cdc56583cd5f39acc7cfd0558f60c668b1f5cbf039b566c1747e12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02c75c8691748a626cbef2de964cb1f4579e702276e8b02d805db0876a8fee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://624c4a09beeecd568d6045a720dfa28c15821d931021cedd8776da4bb36b557f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5b8064433d389b05e3e4744ef4c1f72512db615eae94e2a847bbd4cb2518f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d5b8064433d389b05e3e4744ef4c1f72512db615eae94e2a847bbd4cb2518f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T04:38:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:40:04Z is after 2025-08-24T17:21:41Z" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.920396 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:40:04Z is after 2025-08-24T17:21:41Z" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.931599 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7lb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55156232-4c80-45a7-aad5-a5fa2793e4ed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6e5473c80737a5b7b00e5874062e0485ed7a4d74632202090a9b12f331953b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d2dtw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7lb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:40:04Z is after 2025-08-24T17:21:41Z" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.943497 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.943532 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.943542 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.943560 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.943569 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:04Z","lastTransitionTime":"2025-11-29T04:40:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.943721 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0847696a7b7926047dc62d75739b3b6a7a04cd58efd1c1c2c992d5ae8d20ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83b2ec8f0d384066ba90eb42c1a0a925ade3722b2515b0592d6f2a8f5e533a0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:39:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v22j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-26t88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:40:04Z is after 2025-08-24T17:21:41Z" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.957856 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-qx9wn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce8ce318-27e6-434d-96c8-591d250262fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c99mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c99mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:39:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-qx9wn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:40:04Z is after 2025-08-24T17:21:41Z" Nov 29 04:40:04 crc kubenswrapper[4799]: I1129 04:40:04.973522 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22ed28a-3748-4eb9-b3a7-c02aa38f7e94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:39:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T04:38:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33c0540ccb097b15f65369ff537aef4ae621cc828ba314eb61f0a0037608cadf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c241d29297a73dce12bd0bc2863c56475eb35b8fe8c93991ba0ef8695e38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f42a55d797c6c94af3a5eac1bf08aa38a82359b7ddd0ba32dcc797218ceee6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebed9ec415bb433935da36df8014f9087236cea9788b2048cbff8714f8ad0df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T04:38:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T04:38:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T04:40:04Z is after 2025-08-24T17:21:41Z" Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.047517 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.047604 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.047621 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.047677 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.047694 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:05Z","lastTransitionTime":"2025-11-29T04:40:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.150682 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.150726 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.150735 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.150752 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.150762 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:05Z","lastTransitionTime":"2025-11-29T04:40:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.253380 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.253432 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.253446 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.253464 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.253477 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:05Z","lastTransitionTime":"2025-11-29T04:40:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.356140 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.356210 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.356219 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.356236 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.356245 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:05Z","lastTransitionTime":"2025-11-29T04:40:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.458716 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.458779 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.458879 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.458913 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.458936 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:05Z","lastTransitionTime":"2025-11-29T04:40:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.561490 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.561546 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.561567 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.561593 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.561611 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:05Z","lastTransitionTime":"2025-11-29T04:40:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.658570 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.658598 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.658619 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.658698 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qx9wn" Nov 29 04:40:05 crc kubenswrapper[4799]: E1129 04:40:05.658866 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 04:40:05 crc kubenswrapper[4799]: E1129 04:40:05.659022 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 04:40:05 crc kubenswrapper[4799]: E1129 04:40:05.659186 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 04:40:05 crc kubenswrapper[4799]: E1129 04:40:05.659329 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qx9wn" podUID="ce8ce318-27e6-434d-96c8-591d250262fd" Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.664812 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.664838 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.664847 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.664863 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.664873 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:05Z","lastTransitionTime":"2025-11-29T04:40:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.768686 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.768756 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.768768 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.768807 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.768824 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:05Z","lastTransitionTime":"2025-11-29T04:40:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.871288 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.871361 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.871377 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.871427 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.871446 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:05Z","lastTransitionTime":"2025-11-29T04:40:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.973416 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.973496 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.973522 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.973554 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:05 crc kubenswrapper[4799]: I1129 04:40:05.973580 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:05Z","lastTransitionTime":"2025-11-29T04:40:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:06 crc kubenswrapper[4799]: I1129 04:40:06.077168 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:06 crc kubenswrapper[4799]: I1129 04:40:06.077226 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:06 crc kubenswrapper[4799]: I1129 04:40:06.077238 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:06 crc kubenswrapper[4799]: I1129 04:40:06.077256 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:06 crc kubenswrapper[4799]: I1129 04:40:06.077268 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:06Z","lastTransitionTime":"2025-11-29T04:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:06 crc kubenswrapper[4799]: I1129 04:40:06.180486 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:06 crc kubenswrapper[4799]: I1129 04:40:06.180557 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:06 crc kubenswrapper[4799]: I1129 04:40:06.180581 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:06 crc kubenswrapper[4799]: I1129 04:40:06.180612 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:06 crc kubenswrapper[4799]: I1129 04:40:06.180636 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:06Z","lastTransitionTime":"2025-11-29T04:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:06 crc kubenswrapper[4799]: I1129 04:40:06.283469 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:06 crc kubenswrapper[4799]: I1129 04:40:06.283518 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:06 crc kubenswrapper[4799]: I1129 04:40:06.283528 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:06 crc kubenswrapper[4799]: I1129 04:40:06.283543 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:06 crc kubenswrapper[4799]: I1129 04:40:06.283557 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:06Z","lastTransitionTime":"2025-11-29T04:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:06 crc kubenswrapper[4799]: I1129 04:40:06.386963 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:06 crc kubenswrapper[4799]: I1129 04:40:06.387013 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:06 crc kubenswrapper[4799]: I1129 04:40:06.387025 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:06 crc kubenswrapper[4799]: I1129 04:40:06.387043 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:06 crc kubenswrapper[4799]: I1129 04:40:06.387058 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:06Z","lastTransitionTime":"2025-11-29T04:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:06 crc kubenswrapper[4799]: I1129 04:40:06.489431 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:06 crc kubenswrapper[4799]: I1129 04:40:06.489465 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:06 crc kubenswrapper[4799]: I1129 04:40:06.489477 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:06 crc kubenswrapper[4799]: I1129 04:40:06.489494 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:06 crc kubenswrapper[4799]: I1129 04:40:06.489508 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:06Z","lastTransitionTime":"2025-11-29T04:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:06 crc kubenswrapper[4799]: I1129 04:40:06.591945 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:06 crc kubenswrapper[4799]: I1129 04:40:06.592019 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:06 crc kubenswrapper[4799]: I1129 04:40:06.592036 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:06 crc kubenswrapper[4799]: I1129 04:40:06.592057 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:06 crc kubenswrapper[4799]: I1129 04:40:06.592073 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:06Z","lastTransitionTime":"2025-11-29T04:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:06 crc kubenswrapper[4799]: I1129 04:40:06.694861 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:06 crc kubenswrapper[4799]: I1129 04:40:06.694908 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:06 crc kubenswrapper[4799]: I1129 04:40:06.694924 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:06 crc kubenswrapper[4799]: I1129 04:40:06.694946 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:06 crc kubenswrapper[4799]: I1129 04:40:06.694961 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:06Z","lastTransitionTime":"2025-11-29T04:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:06 crc kubenswrapper[4799]: I1129 04:40:06.796921 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:06 crc kubenswrapper[4799]: I1129 04:40:06.796967 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:06 crc kubenswrapper[4799]: I1129 04:40:06.796975 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:06 crc kubenswrapper[4799]: I1129 04:40:06.796990 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:06 crc kubenswrapper[4799]: I1129 04:40:06.797002 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:06Z","lastTransitionTime":"2025-11-29T04:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:06 crc kubenswrapper[4799]: I1129 04:40:06.899522 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:06 crc kubenswrapper[4799]: I1129 04:40:06.899558 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:06 crc kubenswrapper[4799]: I1129 04:40:06.899566 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:06 crc kubenswrapper[4799]: I1129 04:40:06.899580 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:06 crc kubenswrapper[4799]: I1129 04:40:06.899589 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:06Z","lastTransitionTime":"2025-11-29T04:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.006053 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.006086 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.006094 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.006108 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.006120 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:07Z","lastTransitionTime":"2025-11-29T04:40:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.108185 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.108243 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.108260 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.108285 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.108302 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:07Z","lastTransitionTime":"2025-11-29T04:40:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.212339 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.212428 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.212448 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.212482 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.212502 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:07Z","lastTransitionTime":"2025-11-29T04:40:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.317133 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.317250 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.317276 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.317316 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.317417 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:07Z","lastTransitionTime":"2025-11-29T04:40:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.420626 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.420711 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.420734 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.420765 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.420825 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:07Z","lastTransitionTime":"2025-11-29T04:40:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.523694 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.523758 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.523776 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.523853 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.523872 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:07Z","lastTransitionTime":"2025-11-29T04:40:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.627271 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.627329 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.627339 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.627352 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.627361 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:07Z","lastTransitionTime":"2025-11-29T04:40:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.658919 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.659046 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.659079 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qx9wn" Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.658989 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 04:40:07 crc kubenswrapper[4799]: E1129 04:40:07.659380 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qx9wn" podUID="ce8ce318-27e6-434d-96c8-591d250262fd" Nov 29 04:40:07 crc kubenswrapper[4799]: E1129 04:40:07.659240 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 04:40:07 crc kubenswrapper[4799]: E1129 04:40:07.659610 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 04:40:07 crc kubenswrapper[4799]: E1129 04:40:07.659849 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.730366 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.730439 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.730456 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.730503 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.730524 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:07Z","lastTransitionTime":"2025-11-29T04:40:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.741156 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 04:40:07 crc kubenswrapper[4799]: E1129 04:40:07.741366 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 04:41:11.741323789 +0000 UTC m=+147.384254219 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.741446 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.741576 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.741639 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:40:07 crc kubenswrapper[4799]: E1129 04:40:07.741701 4799 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.741727 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 04:40:07 crc kubenswrapper[4799]: E1129 04:40:07.741766 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 29 04:40:07 crc kubenswrapper[4799]: E1129 04:40:07.741786 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 29 04:40:07 crc kubenswrapper[4799]: E1129 04:40:07.741819 4799 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 04:40:07 crc kubenswrapper[4799]: E1129 04:40:07.741848 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-29 04:41:11.741777551 +0000 UTC m=+147.384708131 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 29 04:40:07 crc kubenswrapper[4799]: E1129 04:40:07.741890 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-29 04:41:11.741871133 +0000 UTC m=+147.384801573 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 04:40:07 crc kubenswrapper[4799]: E1129 04:40:07.741957 4799 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 29 04:40:07 crc kubenswrapper[4799]: E1129 04:40:07.742002 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 29 04:40:07 crc kubenswrapper[4799]: E1129 04:40:07.742025 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-29 04:41:11.742011097 +0000 UTC m=+147.384941507 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 29 04:40:07 crc kubenswrapper[4799]: E1129 04:40:07.742030 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 29 04:40:07 crc kubenswrapper[4799]: E1129 04:40:07.742053 4799 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 04:40:07 crc kubenswrapper[4799]: E1129 04:40:07.742124 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-29 04:41:11.742108679 +0000 UTC m=+147.385039109 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.834725 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.834779 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.834814 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.834841 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.834855 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:07Z","lastTransitionTime":"2025-11-29T04:40:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.938614 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.938683 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.938701 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.938728 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:07 crc kubenswrapper[4799]: I1129 04:40:07.938749 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:07Z","lastTransitionTime":"2025-11-29T04:40:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.041659 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.041737 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.041757 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.041785 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.041842 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:08Z","lastTransitionTime":"2025-11-29T04:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.144588 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.144672 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.144693 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.144721 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.144743 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:08Z","lastTransitionTime":"2025-11-29T04:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.247609 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.247705 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.247732 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.247765 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.247784 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:08Z","lastTransitionTime":"2025-11-29T04:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.350452 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.350527 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.350549 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.350653 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.350680 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:08Z","lastTransitionTime":"2025-11-29T04:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.454271 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.454336 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.454347 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.454383 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.454394 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:08Z","lastTransitionTime":"2025-11-29T04:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.557153 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.557216 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.557238 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.557269 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.557290 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:08Z","lastTransitionTime":"2025-11-29T04:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.659065 4799 scope.go:117] "RemoveContainer" containerID="d5ce3c68997bc5859df929b844d823adf00352454bb278d57e90edcc44817cbe" Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.660814 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.660838 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.660846 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.660859 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.660869 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:08Z","lastTransitionTime":"2025-11-29T04:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.764165 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.764280 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.764298 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.764324 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.764343 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:08Z","lastTransitionTime":"2025-11-29T04:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.866690 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.866743 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.866758 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.866781 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.866831 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:08Z","lastTransitionTime":"2025-11-29T04:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.969315 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.969351 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.969359 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.969374 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:08 crc kubenswrapper[4799]: I1129 04:40:08.969383 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:08Z","lastTransitionTime":"2025-11-29T04:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:09 crc kubenswrapper[4799]: I1129 04:40:09.072850 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:09 crc kubenswrapper[4799]: I1129 04:40:09.072900 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:09 crc kubenswrapper[4799]: I1129 04:40:09.072913 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:09 crc kubenswrapper[4799]: I1129 04:40:09.072930 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:09 crc kubenswrapper[4799]: I1129 04:40:09.072941 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:09Z","lastTransitionTime":"2025-11-29T04:40:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:09 crc kubenswrapper[4799]: I1129 04:40:09.181071 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:09 crc kubenswrapper[4799]: I1129 04:40:09.181112 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:09 crc kubenswrapper[4799]: I1129 04:40:09.181125 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:09 crc kubenswrapper[4799]: I1129 04:40:09.181142 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:09 crc kubenswrapper[4799]: I1129 04:40:09.181155 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:09Z","lastTransitionTime":"2025-11-29T04:40:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:09 crc kubenswrapper[4799]: I1129 04:40:09.240154 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 04:40:09 crc kubenswrapper[4799]: I1129 04:40:09.240206 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 04:40:09 crc kubenswrapper[4799]: I1129 04:40:09.240217 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 04:40:09 crc kubenswrapper[4799]: I1129 04:40:09.240231 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 04:40:09 crc kubenswrapper[4799]: I1129 04:40:09.240241 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T04:40:09Z","lastTransitionTime":"2025-11-29T04:40:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 04:40:09 crc kubenswrapper[4799]: I1129 04:40:09.365053 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-6r2h5"] Nov 29 04:40:09 crc kubenswrapper[4799]: I1129 04:40:09.365941 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6r2h5" Nov 29 04:40:09 crc kubenswrapper[4799]: I1129 04:40:09.368934 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 29 04:40:09 crc kubenswrapper[4799]: I1129 04:40:09.369722 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 29 04:40:09 crc kubenswrapper[4799]: I1129 04:40:09.369872 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 29 04:40:09 crc kubenswrapper[4799]: I1129 04:40:09.370128 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 29 04:40:09 crc kubenswrapper[4799]: I1129 04:40:09.462495 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/edd5841b-6b7b-4128-a55f-7edbf97814d7-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-6r2h5\" (UID: \"edd5841b-6b7b-4128-a55f-7edbf97814d7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6r2h5" Nov 29 04:40:09 crc kubenswrapper[4799]: I1129 04:40:09.462678 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/edd5841b-6b7b-4128-a55f-7edbf97814d7-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-6r2h5\" (UID: \"edd5841b-6b7b-4128-a55f-7edbf97814d7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6r2h5" Nov 29 04:40:09 crc kubenswrapper[4799]: I1129 04:40:09.462716 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/edd5841b-6b7b-4128-a55f-7edbf97814d7-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-6r2h5\" (UID: \"edd5841b-6b7b-4128-a55f-7edbf97814d7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6r2h5" Nov 29 04:40:09 crc kubenswrapper[4799]: I1129 04:40:09.462756 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/edd5841b-6b7b-4128-a55f-7edbf97814d7-service-ca\") pod \"cluster-version-operator-5c965bbfc6-6r2h5\" (UID: \"edd5841b-6b7b-4128-a55f-7edbf97814d7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6r2h5" Nov 29 04:40:09 crc kubenswrapper[4799]: I1129 04:40:09.462864 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/edd5841b-6b7b-4128-a55f-7edbf97814d7-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-6r2h5\" (UID: \"edd5841b-6b7b-4128-a55f-7edbf97814d7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6r2h5" Nov 29 04:40:09 crc kubenswrapper[4799]: I1129 04:40:09.478673 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=37.478647656 podStartE2EDuration="37.478647656s" podCreationTimestamp="2025-11-29 04:39:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:09.439883197 +0000 UTC m=+85.082813647" watchObservedRunningTime="2025-11-29 04:40:09.478647656 +0000 UTC m=+85.121578056" Nov 29 04:40:09 crc kubenswrapper[4799]: I1129 04:40:09.492505 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-v7lb4" podStartSLOduration=60.492469062 podStartE2EDuration="1m0.492469062s" podCreationTimestamp="2025-11-29 04:39:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:09.492107993 +0000 UTC m=+85.135038403" watchObservedRunningTime="2025-11-29 04:40:09.492469062 +0000 UTC m=+85.135399502" Nov 29 04:40:09 crc kubenswrapper[4799]: I1129 04:40:09.509046 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podStartSLOduration=60.509014769 podStartE2EDuration="1m0.509014769s" podCreationTimestamp="2025-11-29 04:39:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:09.508427064 +0000 UTC m=+85.151357504" watchObservedRunningTime="2025-11-29 04:40:09.509014769 +0000 UTC m=+85.151945209" Nov 29 04:40:09 crc kubenswrapper[4799]: I1129 04:40:09.558481 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=61.558451643 podStartE2EDuration="1m1.558451643s" podCreationTimestamp="2025-11-29 04:39:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:09.539934766 +0000 UTC m=+85.182865156" watchObservedRunningTime="2025-11-29 04:40:09.558451643 +0000 UTC m=+85.201382043" Nov 29 04:40:09 crc kubenswrapper[4799]: I1129 04:40:09.563344 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/edd5841b-6b7b-4128-a55f-7edbf97814d7-service-ca\") pod \"cluster-version-operator-5c965bbfc6-6r2h5\" (UID: \"edd5841b-6b7b-4128-a55f-7edbf97814d7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6r2h5" Nov 29 04:40:09 crc kubenswrapper[4799]: I1129 04:40:09.563391 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/edd5841b-6b7b-4128-a55f-7edbf97814d7-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-6r2h5\" (UID: \"edd5841b-6b7b-4128-a55f-7edbf97814d7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6r2h5" Nov 29 04:40:09 crc kubenswrapper[4799]: I1129 04:40:09.563414 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/edd5841b-6b7b-4128-a55f-7edbf97814d7-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-6r2h5\" (UID: \"edd5841b-6b7b-4128-a55f-7edbf97814d7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6r2h5" Nov 29 04:40:09 crc kubenswrapper[4799]: I1129 04:40:09.563443 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/edd5841b-6b7b-4128-a55f-7edbf97814d7-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-6r2h5\" (UID: \"edd5841b-6b7b-4128-a55f-7edbf97814d7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6r2h5" Nov 29 04:40:09 crc kubenswrapper[4799]: I1129 04:40:09.563479 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/edd5841b-6b7b-4128-a55f-7edbf97814d7-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-6r2h5\" (UID: \"edd5841b-6b7b-4128-a55f-7edbf97814d7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6r2h5" Nov 29 04:40:09 crc kubenswrapper[4799]: I1129 04:40:09.563510 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/edd5841b-6b7b-4128-a55f-7edbf97814d7-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-6r2h5\" (UID: \"edd5841b-6b7b-4128-a55f-7edbf97814d7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6r2h5" Nov 29 04:40:09 crc kubenswrapper[4799]: I1129 04:40:09.563581 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/edd5841b-6b7b-4128-a55f-7edbf97814d7-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-6r2h5\" (UID: \"edd5841b-6b7b-4128-a55f-7edbf97814d7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6r2h5" Nov 29 04:40:09 crc kubenswrapper[4799]: I1129 04:40:09.564473 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/edd5841b-6b7b-4128-a55f-7edbf97814d7-service-ca\") pod \"cluster-version-operator-5c965bbfc6-6r2h5\" (UID: \"edd5841b-6b7b-4128-a55f-7edbf97814d7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6r2h5" Nov 29 04:40:09 crc kubenswrapper[4799]: I1129 04:40:09.578207 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/edd5841b-6b7b-4128-a55f-7edbf97814d7-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-6r2h5\" (UID: \"edd5841b-6b7b-4128-a55f-7edbf97814d7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6r2h5" Nov 29 04:40:09 crc kubenswrapper[4799]: I1129 04:40:09.583003 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/edd5841b-6b7b-4128-a55f-7edbf97814d7-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-6r2h5\" (UID: \"edd5841b-6b7b-4128-a55f-7edbf97814d7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6r2h5" Nov 29 04:40:09 crc kubenswrapper[4799]: I1129 04:40:09.643317 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-qt2dl" podStartSLOduration=60.643300061 podStartE2EDuration="1m0.643300061s" podCreationTimestamp="2025-11-29 04:39:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:09.619319213 +0000 UTC m=+85.262249613" watchObservedRunningTime="2025-11-29 04:40:09.643300061 +0000 UTC m=+85.286230461" Nov 29 04:40:09 crc kubenswrapper[4799]: I1129 04:40:09.658052 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 04:40:09 crc kubenswrapper[4799]: I1129 04:40:09.658141 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qx9wn" Nov 29 04:40:09 crc kubenswrapper[4799]: I1129 04:40:09.658173 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 04:40:09 crc kubenswrapper[4799]: E1129 04:40:09.658156 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 04:40:09 crc kubenswrapper[4799]: I1129 04:40:09.658062 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:40:09 crc kubenswrapper[4799]: E1129 04:40:09.658291 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qx9wn" podUID="ce8ce318-27e6-434d-96c8-591d250262fd" Nov 29 04:40:09 crc kubenswrapper[4799]: E1129 04:40:09.658369 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 04:40:09 crc kubenswrapper[4799]: E1129 04:40:09.658424 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 04:40:09 crc kubenswrapper[4799]: I1129 04:40:09.659503 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-s5d7s" podStartSLOduration=60.659493638 podStartE2EDuration="1m0.659493638s" podCreationTimestamp="2025-11-29 04:39:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:09.643466665 +0000 UTC m=+85.286397085" watchObservedRunningTime="2025-11-29 04:40:09.659493638 +0000 UTC m=+85.302424038" Nov 29 04:40:09 crc kubenswrapper[4799]: I1129 04:40:09.659611 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-d87rr" podStartSLOduration=60.659608332 podStartE2EDuration="1m0.659608332s" podCreationTimestamp="2025-11-29 04:39:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:09.657026085 +0000 UTC m=+85.299956495" watchObservedRunningTime="2025-11-29 04:40:09.659608332 +0000 UTC m=+85.302538732" Nov 29 04:40:09 crc kubenswrapper[4799]: I1129 04:40:09.682128 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6r2h5" Nov 29 04:40:09 crc kubenswrapper[4799]: W1129 04:40:09.694727 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podedd5841b_6b7b_4128_a55f_7edbf97814d7.slice/crio-614f026e6d43cd3a7fa1502866045cf043c5dd8c073e8926b21beb94dbcf97b9 WatchSource:0}: Error finding container 614f026e6d43cd3a7fa1502866045cf043c5dd8c073e8926b21beb94dbcf97b9: Status 404 returned error can't find the container with id 614f026e6d43cd3a7fa1502866045cf043c5dd8c073e8926b21beb94dbcf97b9 Nov 29 04:40:09 crc kubenswrapper[4799]: I1129 04:40:09.707961 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v79sj" podStartSLOduration=60.707943377 podStartE2EDuration="1m0.707943377s" podCreationTimestamp="2025-11-29 04:39:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:09.706248534 +0000 UTC m=+85.349178964" watchObservedRunningTime="2025-11-29 04:40:09.707943377 +0000 UTC m=+85.350873777" Nov 29 04:40:10 crc kubenswrapper[4799]: I1129 04:40:10.130858 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=67.130840199 podStartE2EDuration="1m7.130840199s" podCreationTimestamp="2025-11-29 04:39:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:09.736343479 +0000 UTC m=+85.379273879" watchObservedRunningTime="2025-11-29 04:40:10.130840199 +0000 UTC m=+85.773770599" Nov 29 04:40:10 crc kubenswrapper[4799]: I1129 04:40:10.132014 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-qx9wn"] Nov 29 04:40:10 crc kubenswrapper[4799]: I1129 04:40:10.264897 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fht8l_bd007c61-4689-4179-95ad-44a9724e8e7e/ovnkube-controller/2.log" Nov 29 04:40:10 crc kubenswrapper[4799]: I1129 04:40:10.268918 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" event={"ID":"bd007c61-4689-4179-95ad-44a9724e8e7e","Type":"ContainerStarted","Data":"66dd6f66933db1fc65c8aa5ac9ce4117213e4a0a3f1260975e0d9a6e06366f3d"} Nov 29 04:40:10 crc kubenswrapper[4799]: I1129 04:40:10.269592 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:40:10 crc kubenswrapper[4799]: I1129 04:40:10.270761 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qx9wn" Nov 29 04:40:10 crc kubenswrapper[4799]: E1129 04:40:10.270891 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qx9wn" podUID="ce8ce318-27e6-434d-96c8-591d250262fd" Nov 29 04:40:10 crc kubenswrapper[4799]: I1129 04:40:10.271072 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6r2h5" event={"ID":"edd5841b-6b7b-4128-a55f-7edbf97814d7","Type":"ContainerStarted","Data":"c4c0bc80f3e49d846cea801f62bfc9844db6a3a63892287f55fcec8c7bb933a6"} Nov 29 04:40:10 crc kubenswrapper[4799]: I1129 04:40:10.271097 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6r2h5" event={"ID":"edd5841b-6b7b-4128-a55f-7edbf97814d7","Type":"ContainerStarted","Data":"614f026e6d43cd3a7fa1502866045cf043c5dd8c073e8926b21beb94dbcf97b9"} Nov 29 04:40:10 crc kubenswrapper[4799]: I1129 04:40:10.301202 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" podStartSLOduration=61.301183671 podStartE2EDuration="1m1.301183671s" podCreationTimestamp="2025-11-29 04:39:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:10.300019961 +0000 UTC m=+85.942950401" watchObservedRunningTime="2025-11-29 04:40:10.301183671 +0000 UTC m=+85.944114081" Nov 29 04:40:10 crc kubenswrapper[4799]: I1129 04:40:10.317127 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6r2h5" podStartSLOduration=61.317109231 podStartE2EDuration="1m1.317109231s" podCreationTimestamp="2025-11-29 04:39:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:10.315563201 +0000 UTC m=+85.958493611" watchObservedRunningTime="2025-11-29 04:40:10.317109231 +0000 UTC m=+85.960039641" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.658749 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.658762 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.658764 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:40:11 crc kubenswrapper[4799]: E1129 04:40:11.658862 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.658925 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qx9wn" Nov 29 04:40:11 crc kubenswrapper[4799]: E1129 04:40:11.660156 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 04:40:11 crc kubenswrapper[4799]: E1129 04:40:11.660507 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 04:40:11 crc kubenswrapper[4799]: E1129 04:40:11.663090 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qx9wn" podUID="ce8ce318-27e6-434d-96c8-591d250262fd" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.808435 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.808851 4799 kubelet_node_status.go:538] "Fast updating node status as it just became ready" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.845378 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-wznnv"] Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.847441 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-wznnv" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.848639 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-ddl6h"] Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.849164 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-ddl6h" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.849521 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-rcsd7"] Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.849763 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.850225 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-rcsd7" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.851204 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-zv6wk"] Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.851696 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-svdzq"] Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.852033 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-svdzq" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.852458 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zv6wk" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.852758 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.853156 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.853280 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.853508 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.854579 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.854574 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.854730 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.855062 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.855168 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.855179 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.855373 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.855501 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.855612 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.855687 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.855769 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.855775 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.857472 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.857524 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.857576 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.857582 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.857661 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.857739 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.858039 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.858269 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.858442 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.858635 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.858722 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.858727 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.858825 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.859043 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.860640 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.860645 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.861182 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.862635 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-z7df5"] Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.863219 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-z7df5" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.867522 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-6jsdb"] Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.868121 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6jsdb" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.868162 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kp5gm"] Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.868868 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kp5gm" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.874060 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.876204 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.876629 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.877007 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.878085 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.878273 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.878519 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.878728 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.878723 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-tqr7m"] Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.878955 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.879165 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.879397 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.879617 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.879669 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vk2xx"] Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.880051 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-tqr7m" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.880762 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-bpxtd"] Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.886389 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vk2xx" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.886970 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/daa3f105-2960-45a0-ab05-cb2cedea7e9f-images\") pod \"machine-api-operator-5694c8668f-wznnv\" (UID: \"daa3f105-2960-45a0-ab05-cb2cedea7e9f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wznnv" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.887018 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/daa3f105-2960-45a0-ab05-cb2cedea7e9f-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-wznnv\" (UID: \"daa3f105-2960-45a0-ab05-cb2cedea7e9f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wznnv" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.887073 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/daa3f105-2960-45a0-ab05-cb2cedea7e9f-config\") pod \"machine-api-operator-5694c8668f-wznnv\" (UID: \"daa3f105-2960-45a0-ab05-cb2cedea7e9f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wznnv" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.887121 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6nf4\" (UniqueName: \"kubernetes.io/projected/daa3f105-2960-45a0-ab05-cb2cedea7e9f-kube-api-access-g6nf4\") pod \"machine-api-operator-5694c8668f-wznnv\" (UID: \"daa3f105-2960-45a0-ab05-cb2cedea7e9f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wznnv" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.889150 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.889390 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.889560 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.889448 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.889943 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.893628 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-fnhgz"] Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.894113 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-bpxtd" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.894269 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.924636 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.925178 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.927122 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.927729 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.927950 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.928077 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.928278 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.928400 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.928588 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.928775 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.929073 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.929700 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.929831 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.929988 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.930277 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.930391 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-hkpdn"] Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.932084 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-fnhgz" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.932698 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.936552 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-2qb64"] Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.937034 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-vdxqp"] Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.937347 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4bfz8"] Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.937631 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-brs97"] Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.937928 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-bbm46"] Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.938267 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-brs97" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.938326 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-2qb64" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.938652 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-hkpdn" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.938669 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.938736 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4bfz8" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.938740 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.939186 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xj4jw"] Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.939780 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-bbm46" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.941884 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-jrhwp"] Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.942301 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.942399 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-85llw"] Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.942447 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.942562 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.942876 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-85llw" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.943015 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.943094 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xj4jw" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.943304 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-jrhwp" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.943682 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.943912 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.944064 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.944177 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.946163 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-6gm29"] Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.946918 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-9j7n6"] Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.947410 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6gm29" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.951094 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.952591 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.952859 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.952993 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.953123 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.953393 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.953523 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.953640 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.953923 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.954067 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.954193 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.954431 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.954626 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.954754 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.954827 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.954864 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.954934 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.955228 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.955375 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.955390 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.955543 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.957246 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.957829 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9j7n6" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.971197 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jq22x"] Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.972440 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qz5br"] Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.973829 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-t4f4r"] Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.978471 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jq22x" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.981328 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qz5br" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.998210 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.999162 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hs96x"] Nov 29 04:40:11 crc kubenswrapper[4799]: I1129 04:40:11.999619 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hs96x" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.001399 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8j7sv\" (UniqueName: \"kubernetes.io/projected/1c7126e3-85fe-4158-b92c-df2eb2717217-kube-api-access-8j7sv\") pod \"controller-manager-879f6c89f-ddl6h\" (UID: \"1c7126e3-85fe-4158-b92c-df2eb2717217\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ddl6h" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.001450 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-vdxqp\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.001480 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-vdxqp\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.001501 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7pq8\" (UniqueName: \"kubernetes.io/projected/36a3c13c-2670-4b33-a9fe-9071b35365fe-kube-api-access-r7pq8\") pod \"router-default-5444994796-jrhwp\" (UID: \"36a3c13c-2670-4b33-a9fe-9071b35365fe\") " pod="openshift-ingress/router-default-5444994796-jrhwp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.001520 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-audit-dir\") pod \"oauth-openshift-558db77b4-vdxqp\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.001539 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rj74b\" (UniqueName: \"kubernetes.io/projected/b6794cf7-9161-499d-ba18-74b95caaf3ae-kube-api-access-rj74b\") pod \"console-f9d7485db-2qb64\" (UID: \"b6794cf7-9161-499d-ba18-74b95caaf3ae\") " pod="openshift-console/console-f9d7485db-2qb64" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.001556 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mctqc\" (UniqueName: \"kubernetes.io/projected/50764326-a80d-4b7a-bfd3-2242f4d0e228-kube-api-access-mctqc\") pod \"authentication-operator-69f744f599-fnhgz\" (UID: \"50764326-a80d-4b7a-bfd3-2242f4d0e228\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fnhgz" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.001583 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/938dcbc0-ed30-4200-9f6f-cbc1364cce88-trusted-ca\") pod \"console-operator-58897d9998-svdzq\" (UID: \"938dcbc0-ed30-4200-9f6f-cbc1364cce88\") " pod="openshift-console-operator/console-operator-58897d9998-svdzq" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.001599 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c7126e3-85fe-4158-b92c-df2eb2717217-config\") pod \"controller-manager-879f6c89f-ddl6h\" (UID: \"1c7126e3-85fe-4158-b92c-df2eb2717217\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ddl6h" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.001614 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a98eb357-7807-407c-a1e0-9d9b2b94e1fa-config\") pod \"route-controller-manager-6576b87f9c-z7df5\" (UID: \"a98eb357-7807-407c-a1e0-9d9b2b94e1fa\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-z7df5" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.001632 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/938dcbc0-ed30-4200-9f6f-cbc1364cce88-config\") pod \"console-operator-58897d9998-svdzq\" (UID: \"938dcbc0-ed30-4200-9f6f-cbc1364cce88\") " pod="openshift-console-operator/console-operator-58897d9998-svdzq" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.001647 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/9f2aa3db-bf5b-47b3-bcec-5132916f0f7d-etcd-service-ca\") pod \"etcd-operator-b45778765-hkpdn\" (UID: \"9f2aa3db-bf5b-47b3-bcec-5132916f0f7d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hkpdn" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.001664 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/34ecc5fd-acf1-4748-af5d-d9af38bff6d3-audit\") pod \"apiserver-76f77b778f-rcsd7\" (UID: \"34ecc5fd-acf1-4748-af5d-d9af38bff6d3\") " pod="openshift-apiserver/apiserver-76f77b778f-rcsd7" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.001681 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/191743db-39d2-4f93-add6-471de6ccdede-serving-cert\") pod \"apiserver-7bbb656c7d-zv6wk\" (UID: \"191743db-39d2-4f93-add6-471de6ccdede\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zv6wk" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.001703 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/19582088-032e-4b42-bc19-5080f7c97e13-trusted-ca\") pod \"ingress-operator-5b745b69d9-9j7n6\" (UID: \"19582088-032e-4b42-bc19-5080f7c97e13\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9j7n6" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.001720 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tbbpk\" (UniqueName: \"kubernetes.io/projected/7e204bf3-844c-460e-9f4a-7d78025a3b09-kube-api-access-tbbpk\") pod \"openshift-apiserver-operator-796bbdcf4f-kp5gm\" (UID: \"7e204bf3-844c-460e-9f4a-7d78025a3b09\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kp5gm" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.001741 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/50764326-a80d-4b7a-bfd3-2242f4d0e228-service-ca-bundle\") pod \"authentication-operator-69f744f599-fnhgz\" (UID: \"50764326-a80d-4b7a-bfd3-2242f4d0e228\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fnhgz" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.001762 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/b22a0046-b9c4-4b6e-ba05-1c453c998ff8-available-featuregates\") pod \"openshift-config-operator-7777fb866f-tqr7m\" (UID: \"b22a0046-b9c4-4b6e-ba05-1c453c998ff8\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-tqr7m" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.001782 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cdqg\" (UniqueName: \"kubernetes.io/projected/19582088-032e-4b42-bc19-5080f7c97e13-kube-api-access-4cdqg\") pod \"ingress-operator-5b745b69d9-9j7n6\" (UID: \"19582088-032e-4b42-bc19-5080f7c97e13\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9j7n6" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.001817 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xb69\" (UniqueName: \"kubernetes.io/projected/ee4ae924-84bb-4b5c-9b6e-f03d6cba2ad3-kube-api-access-8xb69\") pod \"dns-operator-744455d44c-bpxtd\" (UID: \"ee4ae924-84bb-4b5c-9b6e-f03d6cba2ad3\") " pod="openshift-dns-operator/dns-operator-744455d44c-bpxtd" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.002010 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/daa3f105-2960-45a0-ab05-cb2cedea7e9f-images\") pod \"machine-api-operator-5694c8668f-wznnv\" (UID: \"daa3f105-2960-45a0-ab05-cb2cedea7e9f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wznnv" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.002032 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/34ecc5fd-acf1-4748-af5d-d9af38bff6d3-trusted-ca-bundle\") pod \"apiserver-76f77b778f-rcsd7\" (UID: \"34ecc5fd-acf1-4748-af5d-d9af38bff6d3\") " pod="openshift-apiserver/apiserver-76f77b778f-rcsd7" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.002051 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a98eb357-7807-407c-a1e0-9d9b2b94e1fa-client-ca\") pod \"route-controller-manager-6576b87f9c-z7df5\" (UID: \"a98eb357-7807-407c-a1e0-9d9b2b94e1fa\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-z7df5" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.002069 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e4bb8a0e-4090-4de8-b7b9-5b6fdc899cd0-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-4bfz8\" (UID: \"e4bb8a0e-4090-4de8-b7b9-5b6fdc899cd0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4bfz8" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.002088 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-vdxqp\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.002105 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b22a0046-b9c4-4b6e-ba05-1c453c998ff8-serving-cert\") pod \"openshift-config-operator-7777fb866f-tqr7m\" (UID: \"b22a0046-b9c4-4b6e-ba05-1c453c998ff8\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-tqr7m" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.002123 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/36a3c13c-2670-4b33-a9fe-9071b35365fe-stats-auth\") pod \"router-default-5444994796-jrhwp\" (UID: \"36a3c13c-2670-4b33-a9fe-9071b35365fe\") " pod="openshift-ingress/router-default-5444994796-jrhwp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.002141 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-vdxqp\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.002162 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42edbb7d-0d06-4a5f-b481-7345a1b67227-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-brs97\" (UID: \"42edbb7d-0d06-4a5f-b481-7345a1b67227\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-brs97" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.002181 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9f245325-f6b0-4ffa-b423-63a47df33136-auth-proxy-config\") pod \"machine-approver-56656f9798-6jsdb\" (UID: \"9f245325-f6b0-4ffa-b423-63a47df33136\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6jsdb" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.002197 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlnqf\" (UniqueName: \"kubernetes.io/projected/191743db-39d2-4f93-add6-471de6ccdede-kube-api-access-rlnqf\") pod \"apiserver-7bbb656c7d-zv6wk\" (UID: \"191743db-39d2-4f93-add6-471de6ccdede\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zv6wk" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.002217 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/daa3f105-2960-45a0-ab05-cb2cedea7e9f-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-wznnv\" (UID: \"daa3f105-2960-45a0-ab05-cb2cedea7e9f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wznnv" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.002235 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b6794cf7-9161-499d-ba18-74b95caaf3ae-console-serving-cert\") pod \"console-f9d7485db-2qb64\" (UID: \"b6794cf7-9161-499d-ba18-74b95caaf3ae\") " pod="openshift-console/console-f9d7485db-2qb64" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.002255 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e73c6003-61f1-49b4-9275-14a55a35e186-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-6gm29\" (UID: \"e73c6003-61f1-49b4-9275-14a55a35e186\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6gm29" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.002274 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/36a3c13c-2670-4b33-a9fe-9071b35365fe-service-ca-bundle\") pod \"router-default-5444994796-jrhwp\" (UID: \"36a3c13c-2670-4b33-a9fe-9071b35365fe\") " pod="openshift-ingress/router-default-5444994796-jrhwp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.002305 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lczb2\" (UniqueName: \"kubernetes.io/projected/fdba8af3-a255-470f-bffd-ced0a45e988d-kube-api-access-lczb2\") pod \"downloads-7954f5f757-bbm46\" (UID: \"fdba8af3-a255-470f-bffd-ced0a45e988d\") " pod="openshift-console/downloads-7954f5f757-bbm46" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.002330 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-vdxqp\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.002363 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34ecc5fd-acf1-4748-af5d-d9af38bff6d3-config\") pod \"apiserver-76f77b778f-rcsd7\" (UID: \"34ecc5fd-acf1-4748-af5d-d9af38bff6d3\") " pod="openshift-apiserver/apiserver-76f77b778f-rcsd7" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.002383 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfk8z\" (UniqueName: \"kubernetes.io/projected/34ecc5fd-acf1-4748-af5d-d9af38bff6d3-kube-api-access-sfk8z\") pod \"apiserver-76f77b778f-rcsd7\" (UID: \"34ecc5fd-acf1-4748-af5d-d9af38bff6d3\") " pod="openshift-apiserver/apiserver-76f77b778f-rcsd7" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.002445 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/19582088-032e-4b42-bc19-5080f7c97e13-metrics-tls\") pod \"ingress-operator-5b745b69d9-9j7n6\" (UID: \"19582088-032e-4b42-bc19-5080f7c97e13\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9j7n6" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.002737 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-nmvwb"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.002902 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.003298 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/daa3f105-2960-45a0-ab05-cb2cedea7e9f-images\") pod \"machine-api-operator-5694c8668f-wznnv\" (UID: \"daa3f105-2960-45a0-ab05-cb2cedea7e9f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wznnv" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.003362 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/34ecc5fd-acf1-4748-af5d-d9af38bff6d3-audit-dir\") pod \"apiserver-76f77b778f-rcsd7\" (UID: \"34ecc5fd-acf1-4748-af5d-d9af38bff6d3\") " pod="openshift-apiserver/apiserver-76f77b778f-rcsd7" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.003381 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/191743db-39d2-4f93-add6-471de6ccdede-etcd-client\") pod \"apiserver-7bbb656c7d-zv6wk\" (UID: \"191743db-39d2-4f93-add6-471de6ccdede\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zv6wk" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.003398 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/191743db-39d2-4f93-add6-471de6ccdede-audit-dir\") pod \"apiserver-7bbb656c7d-zv6wk\" (UID: \"191743db-39d2-4f93-add6-471de6ccdede\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zv6wk" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.003418 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/19582088-032e-4b42-bc19-5080f7c97e13-bound-sa-token\") pod \"ingress-operator-5b745b69d9-9j7n6\" (UID: \"19582088-032e-4b42-bc19-5080f7c97e13\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9j7n6" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.003439 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-audit-policies\") pod \"oauth-openshift-558db77b4-vdxqp\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.003457 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gc56\" (UniqueName: \"kubernetes.io/projected/42edbb7d-0d06-4a5f-b481-7345a1b67227-kube-api-access-6gc56\") pod \"openshift-controller-manager-operator-756b6f6bc6-brs97\" (UID: \"42edbb7d-0d06-4a5f-b481-7345a1b67227\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-brs97" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.003472 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/34ecc5fd-acf1-4748-af5d-d9af38bff6d3-etcd-serving-ca\") pod \"apiserver-76f77b778f-rcsd7\" (UID: \"34ecc5fd-acf1-4748-af5d-d9af38bff6d3\") " pod="openshift-apiserver/apiserver-76f77b778f-rcsd7" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.003487 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/34ecc5fd-acf1-4748-af5d-d9af38bff6d3-image-import-ca\") pod \"apiserver-76f77b778f-rcsd7\" (UID: \"34ecc5fd-acf1-4748-af5d-d9af38bff6d3\") " pod="openshift-apiserver/apiserver-76f77b778f-rcsd7" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.003502 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/191743db-39d2-4f93-add6-471de6ccdede-encryption-config\") pod \"apiserver-7bbb656c7d-zv6wk\" (UID: \"191743db-39d2-4f93-add6-471de6ccdede\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zv6wk" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.003517 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e204bf3-844c-460e-9f4a-7d78025a3b09-config\") pod \"openshift-apiserver-operator-796bbdcf4f-kp5gm\" (UID: \"7e204bf3-844c-460e-9f4a-7d78025a3b09\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kp5gm" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.003586 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f2aa3db-bf5b-47b3-bcec-5132916f0f7d-config\") pod \"etcd-operator-b45778765-hkpdn\" (UID: \"9f2aa3db-bf5b-47b3-bcec-5132916f0f7d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hkpdn" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.003606 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q67cq\" (UniqueName: \"kubernetes.io/projected/e86c829a-524d-4194-9ab2-717b21d7cd60-kube-api-access-q67cq\") pod \"cluster-samples-operator-665b6dd947-vk2xx\" (UID: \"e86c829a-524d-4194-9ab2-717b21d7cd60\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vk2xx" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.003625 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/bfdd16e4-1cb8-4178-af49-4db764abc507-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-85llw\" (UID: \"bfdd16e4-1cb8-4178-af49-4db764abc507\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-85llw" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.003643 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2kp2\" (UniqueName: \"kubernetes.io/projected/bfdd16e4-1cb8-4178-af49-4db764abc507-kube-api-access-k2kp2\") pod \"control-plane-machine-set-operator-78cbb6b69f-85llw\" (UID: \"bfdd16e4-1cb8-4178-af49-4db764abc507\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-85llw" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.003675 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e73c6003-61f1-49b4-9275-14a55a35e186-proxy-tls\") pod \"machine-config-controller-84d6567774-6gm29\" (UID: \"e73c6003-61f1-49b4-9275-14a55a35e186\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6gm29" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.003692 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-vdxqp\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.003708 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xstcq\" (UniqueName: \"kubernetes.io/projected/b22a0046-b9c4-4b6e-ba05-1c453c998ff8-kube-api-access-xstcq\") pod \"openshift-config-operator-7777fb866f-tqr7m\" (UID: \"b22a0046-b9c4-4b6e-ba05-1c453c998ff8\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-tqr7m" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.003723 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4bb8a0e-4090-4de8-b7b9-5b6fdc899cd0-config\") pod \"kube-apiserver-operator-766d6c64bb-4bfz8\" (UID: \"e4bb8a0e-4090-4de8-b7b9-5b6fdc899cd0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4bfz8" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.003738 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/191743db-39d2-4f93-add6-471de6ccdede-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-zv6wk\" (UID: \"191743db-39d2-4f93-add6-471de6ccdede\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zv6wk" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.003758 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6nf4\" (UniqueName: \"kubernetes.io/projected/daa3f105-2960-45a0-ab05-cb2cedea7e9f-kube-api-access-g6nf4\") pod \"machine-api-operator-5694c8668f-wznnv\" (UID: \"daa3f105-2960-45a0-ab05-cb2cedea7e9f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wznnv" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.003774 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b6794cf7-9161-499d-ba18-74b95caaf3ae-console-config\") pod \"console-f9d7485db-2qb64\" (UID: \"b6794cf7-9161-499d-ba18-74b95caaf3ae\") " pod="openshift-console/console-f9d7485db-2qb64" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.003829 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/938dcbc0-ed30-4200-9f6f-cbc1364cce88-serving-cert\") pod \"console-operator-58897d9998-svdzq\" (UID: \"938dcbc0-ed30-4200-9f6f-cbc1364cce88\") " pod="openshift-console-operator/console-operator-58897d9998-svdzq" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.003856 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1c7126e3-85fe-4158-b92c-df2eb2717217-serving-cert\") pod \"controller-manager-879f6c89f-ddl6h\" (UID: \"1c7126e3-85fe-4158-b92c-df2eb2717217\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ddl6h" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.003874 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-vdxqp\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.003892 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/34ecc5fd-acf1-4748-af5d-d9af38bff6d3-etcd-client\") pod \"apiserver-76f77b778f-rcsd7\" (UID: \"34ecc5fd-acf1-4748-af5d-d9af38bff6d3\") " pod="openshift-apiserver/apiserver-76f77b778f-rcsd7" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.003907 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/50764326-a80d-4b7a-bfd3-2242f4d0e228-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-fnhgz\" (UID: \"50764326-a80d-4b7a-bfd3-2242f4d0e228\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fnhgz" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.003974 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b6794cf7-9161-499d-ba18-74b95caaf3ae-console-oauth-config\") pod \"console-f9d7485db-2qb64\" (UID: \"b6794cf7-9161-499d-ba18-74b95caaf3ae\") " pod="openshift-console/console-f9d7485db-2qb64" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.003990 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1c7126e3-85fe-4158-b92c-df2eb2717217-client-ca\") pod \"controller-manager-879f6c89f-ddl6h\" (UID: \"1c7126e3-85fe-4158-b92c-df2eb2717217\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ddl6h" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.004006 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqwlf\" (UniqueName: \"kubernetes.io/projected/9f245325-f6b0-4ffa-b423-63a47df33136-kube-api-access-mqwlf\") pod \"machine-approver-56656f9798-6jsdb\" (UID: \"9f245325-f6b0-4ffa-b423-63a47df33136\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6jsdb" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.004025 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/34ecc5fd-acf1-4748-af5d-d9af38bff6d3-node-pullsecrets\") pod \"apiserver-76f77b778f-rcsd7\" (UID: \"34ecc5fd-acf1-4748-af5d-d9af38bff6d3\") " pod="openshift-apiserver/apiserver-76f77b778f-rcsd7" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.004043 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/34ecc5fd-acf1-4748-af5d-d9af38bff6d3-encryption-config\") pod \"apiserver-76f77b778f-rcsd7\" (UID: \"34ecc5fd-acf1-4748-af5d-d9af38bff6d3\") " pod="openshift-apiserver/apiserver-76f77b778f-rcsd7" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.004063 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/36a3c13c-2670-4b33-a9fe-9071b35365fe-metrics-certs\") pod \"router-default-5444994796-jrhwp\" (UID: \"36a3c13c-2670-4b33-a9fe-9071b35365fe\") " pod="openshift-ingress/router-default-5444994796-jrhwp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.004077 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a98eb357-7807-407c-a1e0-9d9b2b94e1fa-serving-cert\") pod \"route-controller-manager-6576b87f9c-z7df5\" (UID: \"a98eb357-7807-407c-a1e0-9d9b2b94e1fa\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-z7df5" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.004091 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e4bb8a0e-4090-4de8-b7b9-5b6fdc899cd0-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-4bfz8\" (UID: \"e4bb8a0e-4090-4de8-b7b9-5b6fdc899cd0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4bfz8" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.004115 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b6794cf7-9161-499d-ba18-74b95caaf3ae-oauth-serving-cert\") pod \"console-f9d7485db-2qb64\" (UID: \"b6794cf7-9161-499d-ba18-74b95caaf3ae\") " pod="openshift-console/console-f9d7485db-2qb64" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.004133 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74f1d4e8-2506-43c8-bcbb-cb8bf28cfb35-config\") pod \"kube-controller-manager-operator-78b949d7b-xj4jw\" (UID: \"74f1d4e8-2506-43c8-bcbb-cb8bf28cfb35\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xj4jw" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.004149 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-vdxqp\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.004314 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f2aa3db-bf5b-47b3-bcec-5132916f0f7d-serving-cert\") pod \"etcd-operator-b45778765-hkpdn\" (UID: \"9f2aa3db-bf5b-47b3-bcec-5132916f0f7d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hkpdn" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.004333 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-vdxqp\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.004351 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74f1d4e8-2506-43c8-bcbb-cb8bf28cfb35-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-xj4jw\" (UID: \"74f1d4e8-2506-43c8-bcbb-cb8bf28cfb35\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xj4jw" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.004385 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1c7126e3-85fe-4158-b92c-df2eb2717217-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-ddl6h\" (UID: \"1c7126e3-85fe-4158-b92c-df2eb2717217\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ddl6h" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.004402 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/74f1d4e8-2506-43c8-bcbb-cb8bf28cfb35-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-xj4jw\" (UID: \"74f1d4e8-2506-43c8-bcbb-cb8bf28cfb35\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xj4jw" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.004419 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e204bf3-844c-460e-9f4a-7d78025a3b09-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-kp5gm\" (UID: \"7e204bf3-844c-460e-9f4a-7d78025a3b09\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kp5gm" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.004436 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/50764326-a80d-4b7a-bfd3-2242f4d0e228-config\") pod \"authentication-operator-69f744f599-fnhgz\" (UID: \"50764326-a80d-4b7a-bfd3-2242f4d0e228\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fnhgz" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.004453 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b6794cf7-9161-499d-ba18-74b95caaf3ae-service-ca\") pod \"console-f9d7485db-2qb64\" (UID: \"b6794cf7-9161-499d-ba18-74b95caaf3ae\") " pod="openshift-console/console-f9d7485db-2qb64" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.004469 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b6794cf7-9161-499d-ba18-74b95caaf3ae-trusted-ca-bundle\") pod \"console-f9d7485db-2qb64\" (UID: \"b6794cf7-9161-499d-ba18-74b95caaf3ae\") " pod="openshift-console/console-f9d7485db-2qb64" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.004486 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfljm\" (UniqueName: \"kubernetes.io/projected/e73c6003-61f1-49b4-9275-14a55a35e186-kube-api-access-tfljm\") pod \"machine-config-controller-84d6567774-6gm29\" (UID: \"e73c6003-61f1-49b4-9275-14a55a35e186\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6gm29" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.004505 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9f2aa3db-bf5b-47b3-bcec-5132916f0f7d-etcd-client\") pod \"etcd-operator-b45778765-hkpdn\" (UID: \"9f2aa3db-bf5b-47b3-bcec-5132916f0f7d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hkpdn" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.004549 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nmvwb" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.004564 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-vdxqp\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.004584 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/e86c829a-524d-4194-9ab2-717b21d7cd60-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-vk2xx\" (UID: \"e86c829a-524d-4194-9ab2-717b21d7cd60\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vk2xx" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.004602 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/36a3c13c-2670-4b33-a9fe-9071b35365fe-default-certificate\") pod \"router-default-5444994796-jrhwp\" (UID: \"36a3c13c-2670-4b33-a9fe-9071b35365fe\") " pod="openshift-ingress/router-default-5444994796-jrhwp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.004638 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvdj4\" (UniqueName: \"kubernetes.io/projected/a98eb357-7807-407c-a1e0-9d9b2b94e1fa-kube-api-access-kvdj4\") pod \"route-controller-manager-6576b87f9c-z7df5\" (UID: \"a98eb357-7807-407c-a1e0-9d9b2b94e1fa\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-z7df5" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.004639 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-2bbp8"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.004660 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/daa3f105-2960-45a0-ab05-cb2cedea7e9f-config\") pod \"machine-api-operator-5694c8668f-wznnv\" (UID: \"daa3f105-2960-45a0-ab05-cb2cedea7e9f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wznnv" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.004681 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/9f2aa3db-bf5b-47b3-bcec-5132916f0f7d-etcd-ca\") pod \"etcd-operator-b45778765-hkpdn\" (UID: \"9f2aa3db-bf5b-47b3-bcec-5132916f0f7d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hkpdn" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.004697 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/9f245325-f6b0-4ffa-b423-63a47df33136-machine-approver-tls\") pod \"machine-approver-56656f9798-6jsdb\" (UID: \"9f245325-f6b0-4ffa-b423-63a47df33136\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6jsdb" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.004832 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t4f4r" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.005269 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/daa3f105-2960-45a0-ab05-cb2cedea7e9f-config\") pod \"machine-api-operator-5694c8668f-wznnv\" (UID: \"daa3f105-2960-45a0-ab05-cb2cedea7e9f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wznnv" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.005328 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/191743db-39d2-4f93-add6-471de6ccdede-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-zv6wk\" (UID: \"191743db-39d2-4f93-add6-471de6ccdede\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zv6wk" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.005346 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/50764326-a80d-4b7a-bfd3-2242f4d0e228-serving-cert\") pod \"authentication-operator-69f744f599-fnhgz\" (UID: \"50764326-a80d-4b7a-bfd3-2242f4d0e228\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fnhgz" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.005363 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/34ecc5fd-acf1-4748-af5d-d9af38bff6d3-serving-cert\") pod \"apiserver-76f77b778f-rcsd7\" (UID: \"34ecc5fd-acf1-4748-af5d-d9af38bff6d3\") " pod="openshift-apiserver/apiserver-76f77b778f-rcsd7" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.005379 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/191743db-39d2-4f93-add6-471de6ccdede-audit-policies\") pod \"apiserver-7bbb656c7d-zv6wk\" (UID: \"191743db-39d2-4f93-add6-471de6ccdede\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zv6wk" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.005397 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-vdxqp\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.005437 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6t28\" (UniqueName: \"kubernetes.io/projected/938dcbc0-ed30-4200-9f6f-cbc1364cce88-kube-api-access-v6t28\") pod \"console-operator-58897d9998-svdzq\" (UID: \"938dcbc0-ed30-4200-9f6f-cbc1364cce88\") " pod="openshift-console-operator/console-operator-58897d9998-svdzq" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.005452 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.005454 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vv4zx\" (UniqueName: \"kubernetes.io/projected/9f2aa3db-bf5b-47b3-bcec-5132916f0f7d-kube-api-access-vv4zx\") pod \"etcd-operator-b45778765-hkpdn\" (UID: \"9f2aa3db-bf5b-47b3-bcec-5132916f0f7d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hkpdn" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.005592 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrk57\" (UniqueName: \"kubernetes.io/projected/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-kube-api-access-jrk57\") pod \"oauth-openshift-558db77b4-vdxqp\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.005701 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/42edbb7d-0d06-4a5f-b481-7345a1b67227-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-brs97\" (UID: \"42edbb7d-0d06-4a5f-b481-7345a1b67227\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-brs97" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.005736 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f245325-f6b0-4ffa-b423-63a47df33136-config\") pod \"machine-approver-56656f9798-6jsdb\" (UID: \"9f245325-f6b0-4ffa-b423-63a47df33136\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6jsdb" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.005763 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ee4ae924-84bb-4b5c-9b6e-f03d6cba2ad3-metrics-tls\") pod \"dns-operator-744455d44c-bpxtd\" (UID: \"ee4ae924-84bb-4b5c-9b6e-f03d6cba2ad3\") " pod="openshift-dns-operator/dns-operator-744455d44c-bpxtd" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.006569 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.011524 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.012685 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-2t46v"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.013588 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9krqq"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.013759 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/daa3f105-2960-45a0-ab05-cb2cedea7e9f-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-wznnv\" (UID: \"daa3f105-2960-45a0-ab05-cb2cedea7e9f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wznnv" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.014074 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9krqq" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.014190 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-2t46v" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.015007 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g8b88"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.016622 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-6rp4p"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.016856 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g8b88" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.017157 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-6rp4p" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.017322 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h4w54"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.018075 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h4w54" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.018213 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-wznnv"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.019060 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-5btb4"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.019490 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-5btb4" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.019911 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406510-f2lz7"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.020565 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406510-f2lz7" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.020817 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rhzbg"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.021258 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-rhzbg" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.021557 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qd7d2"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.022422 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qd7d2" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.023388 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-ddl6h"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.023989 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.028374 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kp5gm"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.029462 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-zv6wk"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.029528 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.030598 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-rcsd7"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.034045 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vk2xx"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.035712 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-brs97"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.038109 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-fnhgz"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.040862 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-hkpdn"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.042765 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-85llw"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.053720 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-tqr7m"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.053915 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-bpxtd"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.063215 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.074059 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.082636 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-2t46v"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.105145 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-9j7n6"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.105191 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-vdxqp"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.108221 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f2aa3db-bf5b-47b3-bcec-5132916f0f7d-config\") pod \"etcd-operator-b45778765-hkpdn\" (UID: \"9f2aa3db-bf5b-47b3-bcec-5132916f0f7d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hkpdn" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.108305 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/34ecc5fd-acf1-4748-af5d-d9af38bff6d3-image-import-ca\") pod \"apiserver-76f77b778f-rcsd7\" (UID: \"34ecc5fd-acf1-4748-af5d-d9af38bff6d3\") " pod="openshift-apiserver/apiserver-76f77b778f-rcsd7" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.108378 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q67cq\" (UniqueName: \"kubernetes.io/projected/e86c829a-524d-4194-9ab2-717b21d7cd60-kube-api-access-q67cq\") pod \"cluster-samples-operator-665b6dd947-vk2xx\" (UID: \"e86c829a-524d-4194-9ab2-717b21d7cd60\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vk2xx" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.108431 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/f42f4e15-4186-4a35-aed6-122f4518d222-signing-key\") pod \"service-ca-9c57cc56f-5btb4\" (UID: \"f42f4e15-4186-4a35-aed6-122f4518d222\") " pod="openshift-service-ca/service-ca-9c57cc56f-5btb4" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.108475 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e73c6003-61f1-49b4-9275-14a55a35e186-proxy-tls\") pod \"machine-config-controller-84d6567774-6gm29\" (UID: \"e73c6003-61f1-49b4-9275-14a55a35e186\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6gm29" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.108515 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2kp2\" (UniqueName: \"kubernetes.io/projected/bfdd16e4-1cb8-4178-af49-4db764abc507-kube-api-access-k2kp2\") pod \"control-plane-machine-set-operator-78cbb6b69f-85llw\" (UID: \"bfdd16e4-1cb8-4178-af49-4db764abc507\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-85llw" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.108546 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/6743e3ae-38d5-4ac1-8eea-7f0fbbe282da-tmpfs\") pod \"packageserver-d55dfcdfc-g8b88\" (UID: \"6743e3ae-38d5-4ac1-8eea-7f0fbbe282da\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g8b88" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.108595 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b6794cf7-9161-499d-ba18-74b95caaf3ae-console-config\") pod \"console-f9d7485db-2qb64\" (UID: \"b6794cf7-9161-499d-ba18-74b95caaf3ae\") " pod="openshift-console/console-f9d7485db-2qb64" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.108637 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-vdxqp\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.108677 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/34ecc5fd-acf1-4748-af5d-d9af38bff6d3-etcd-client\") pod \"apiserver-76f77b778f-rcsd7\" (UID: \"34ecc5fd-acf1-4748-af5d-d9af38bff6d3\") " pod="openshift-apiserver/apiserver-76f77b778f-rcsd7" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.108705 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/50764326-a80d-4b7a-bfd3-2242f4d0e228-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-fnhgz\" (UID: \"50764326-a80d-4b7a-bfd3-2242f4d0e228\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fnhgz" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.108728 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/34ecc5fd-acf1-4748-af5d-d9af38bff6d3-encryption-config\") pod \"apiserver-76f77b778f-rcsd7\" (UID: \"34ecc5fd-acf1-4748-af5d-d9af38bff6d3\") " pod="openshift-apiserver/apiserver-76f77b778f-rcsd7" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.108762 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/36a3c13c-2670-4b33-a9fe-9071b35365fe-metrics-certs\") pod \"router-default-5444994796-jrhwp\" (UID: \"36a3c13c-2670-4b33-a9fe-9071b35365fe\") " pod="openshift-ingress/router-default-5444994796-jrhwp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.108812 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b6794cf7-9161-499d-ba18-74b95caaf3ae-oauth-serving-cert\") pod \"console-f9d7485db-2qb64\" (UID: \"b6794cf7-9161-499d-ba18-74b95caaf3ae\") " pod="openshift-console/console-f9d7485db-2qb64" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.108840 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2f1d8a27-12bd-4ac8-b23f-51f9f10fbcbf-srv-cert\") pod \"catalog-operator-68c6474976-qz5br\" (UID: \"2f1d8a27-12bd-4ac8-b23f-51f9f10fbcbf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qz5br" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.108867 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f2aa3db-bf5b-47b3-bcec-5132916f0f7d-serving-cert\") pod \"etcd-operator-b45778765-hkpdn\" (UID: \"9f2aa3db-bf5b-47b3-bcec-5132916f0f7d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hkpdn" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.108897 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/50764326-a80d-4b7a-bfd3-2242f4d0e228-config\") pod \"authentication-operator-69f744f599-fnhgz\" (UID: \"50764326-a80d-4b7a-bfd3-2242f4d0e228\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fnhgz" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.108964 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b6794cf7-9161-499d-ba18-74b95caaf3ae-service-ca\") pod \"console-f9d7485db-2qb64\" (UID: \"b6794cf7-9161-499d-ba18-74b95caaf3ae\") " pod="openshift-console/console-f9d7485db-2qb64" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.108999 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfljm\" (UniqueName: \"kubernetes.io/projected/e73c6003-61f1-49b4-9275-14a55a35e186-kube-api-access-tfljm\") pod \"machine-config-controller-84d6567774-6gm29\" (UID: \"e73c6003-61f1-49b4-9275-14a55a35e186\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6gm29" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.109038 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-vdxqp\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.109099 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/e86c829a-524d-4194-9ab2-717b21d7cd60-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-vk2xx\" (UID: \"e86c829a-524d-4194-9ab2-717b21d7cd60\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vk2xx" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.109125 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/36a3c13c-2670-4b33-a9fe-9071b35365fe-default-certificate\") pod \"router-default-5444994796-jrhwp\" (UID: \"36a3c13c-2670-4b33-a9fe-9071b35365fe\") " pod="openshift-ingress/router-default-5444994796-jrhwp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.109169 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/96a1e39b-0018-41b8-8d7c-9fa7697858e9-config-volume\") pod \"collect-profiles-29406510-f2lz7\" (UID: \"96a1e39b-0018-41b8-8d7c-9fa7697858e9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406510-f2lz7" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.109314 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/50764326-a80d-4b7a-bfd3-2242f4d0e228-serving-cert\") pod \"authentication-operator-69f744f599-fnhgz\" (UID: \"50764326-a80d-4b7a-bfd3-2242f4d0e228\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fnhgz" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.109342 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/191743db-39d2-4f93-add6-471de6ccdede-audit-policies\") pod \"apiserver-7bbb656c7d-zv6wk\" (UID: \"191743db-39d2-4f93-add6-471de6ccdede\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zv6wk" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.109370 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/42edbb7d-0d06-4a5f-b481-7345a1b67227-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-brs97\" (UID: \"42edbb7d-0d06-4a5f-b481-7345a1b67227\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-brs97" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.109395 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ee4ae924-84bb-4b5c-9b6e-f03d6cba2ad3-metrics-tls\") pod \"dns-operator-744455d44c-bpxtd\" (UID: \"ee4ae924-84bb-4b5c-9b6e-f03d6cba2ad3\") " pod="openshift-dns-operator/dns-operator-744455d44c-bpxtd" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.109431 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrk57\" (UniqueName: \"kubernetes.io/projected/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-kube-api-access-jrk57\") pod \"oauth-openshift-558db77b4-vdxqp\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.109456 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8j7sv\" (UniqueName: \"kubernetes.io/projected/1c7126e3-85fe-4158-b92c-df2eb2717217-kube-api-access-8j7sv\") pod \"controller-manager-879f6c89f-ddl6h\" (UID: \"1c7126e3-85fe-4158-b92c-df2eb2717217\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ddl6h" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.109479 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-vdxqp\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.109559 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nstwg\" (UniqueName: \"kubernetes.io/projected/6743e3ae-38d5-4ac1-8eea-7f0fbbe282da-kube-api-access-nstwg\") pod \"packageserver-d55dfcdfc-g8b88\" (UID: \"6743e3ae-38d5-4ac1-8eea-7f0fbbe282da\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g8b88" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.109608 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rj74b\" (UniqueName: \"kubernetes.io/projected/b6794cf7-9161-499d-ba18-74b95caaf3ae-kube-api-access-rj74b\") pod \"console-f9d7485db-2qb64\" (UID: \"b6794cf7-9161-499d-ba18-74b95caaf3ae\") " pod="openshift-console/console-f9d7485db-2qb64" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.109674 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e69469d9-e3b1-4afd-84de-9b7dafa8e9e4-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-2t46v\" (UID: \"e69469d9-e3b1-4afd-84de-9b7dafa8e9e4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-2t46v" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.109825 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzpbl\" (UniqueName: \"kubernetes.io/projected/fecaea20-4e1a-4977-92e0-d0bd7c3037a2-kube-api-access-fzpbl\") pod \"marketplace-operator-79b997595-rhzbg\" (UID: \"fecaea20-4e1a-4977-92e0-d0bd7c3037a2\") " pod="openshift-marketplace/marketplace-operator-79b997595-rhzbg" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.109876 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/9f2aa3db-bf5b-47b3-bcec-5132916f0f7d-etcd-service-ca\") pod \"etcd-operator-b45778765-hkpdn\" (UID: \"9f2aa3db-bf5b-47b3-bcec-5132916f0f7d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hkpdn" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.109961 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/19582088-032e-4b42-bc19-5080f7c97e13-trusted-ca\") pod \"ingress-operator-5b745b69d9-9j7n6\" (UID: \"19582088-032e-4b42-bc19-5080f7c97e13\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9j7n6" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.110218 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aab927aa-d4cb-4b96-a5a6-bb14df8d4af8-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-hs96x\" (UID: \"aab927aa-d4cb-4b96-a5a6-bb14df8d4af8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hs96x" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.110369 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4cdqg\" (UniqueName: \"kubernetes.io/projected/19582088-032e-4b42-bc19-5080f7c97e13-kube-api-access-4cdqg\") pod \"ingress-operator-5b745b69d9-9j7n6\" (UID: \"19582088-032e-4b42-bc19-5080f7c97e13\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9j7n6" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.110412 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/aab927aa-d4cb-4b96-a5a6-bb14df8d4af8-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-hs96x\" (UID: \"aab927aa-d4cb-4b96-a5a6-bb14df8d4af8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hs96x" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.110549 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f2baa9d1-df52-44c6-b62f-f39c16b70e34-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-jq22x\" (UID: \"f2baa9d1-df52-44c6-b62f-f39c16b70e34\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jq22x" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.110581 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b22a0046-b9c4-4b6e-ba05-1c453c998ff8-serving-cert\") pod \"openshift-config-operator-7777fb866f-tqr7m\" (UID: \"b22a0046-b9c4-4b6e-ba05-1c453c998ff8\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-tqr7m" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.110607 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a98eb357-7807-407c-a1e0-9d9b2b94e1fa-client-ca\") pod \"route-controller-manager-6576b87f9c-z7df5\" (UID: \"a98eb357-7807-407c-a1e0-9d9b2b94e1fa\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-z7df5" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.110669 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvj24\" (UniqueName: \"kubernetes.io/projected/cddf4720-7bab-43e0-989e-6cff2942a5fe-kube-api-access-nvj24\") pod \"package-server-manager-789f6589d5-qd7d2\" (UID: \"cddf4720-7bab-43e0-989e-6cff2942a5fe\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qd7d2" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.110692 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42edbb7d-0d06-4a5f-b481-7345a1b67227-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-brs97\" (UID: \"42edbb7d-0d06-4a5f-b481-7345a1b67227\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-brs97" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.110712 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/96a1e39b-0018-41b8-8d7c-9fa7697858e9-secret-volume\") pod \"collect-profiles-29406510-f2lz7\" (UID: \"96a1e39b-0018-41b8-8d7c-9fa7697858e9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406510-f2lz7" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.110766 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e73c6003-61f1-49b4-9275-14a55a35e186-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-6gm29\" (UID: \"e73c6003-61f1-49b4-9275-14a55a35e186\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6gm29" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.110919 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lczb2\" (UniqueName: \"kubernetes.io/projected/fdba8af3-a255-470f-bffd-ced0a45e988d-kube-api-access-lczb2\") pod \"downloads-7954f5f757-bbm46\" (UID: \"fdba8af3-a255-470f-bffd-ced0a45e988d\") " pod="openshift-console/downloads-7954f5f757-bbm46" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.110995 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/19582088-032e-4b42-bc19-5080f7c97e13-bound-sa-token\") pod \"ingress-operator-5b745b69d9-9j7n6\" (UID: \"19582088-032e-4b42-bc19-5080f7c97e13\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9j7n6" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.111089 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aab927aa-d4cb-4b96-a5a6-bb14df8d4af8-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-hs96x\" (UID: \"aab927aa-d4cb-4b96-a5a6-bb14df8d4af8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hs96x" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.111317 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gc56\" (UniqueName: \"kubernetes.io/projected/42edbb7d-0d06-4a5f-b481-7345a1b67227-kube-api-access-6gc56\") pod \"openshift-controller-manager-operator-756b6f6bc6-brs97\" (UID: \"42edbb7d-0d06-4a5f-b481-7345a1b67227\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-brs97" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.111336 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/34ecc5fd-acf1-4748-af5d-d9af38bff6d3-etcd-serving-ca\") pod \"apiserver-76f77b778f-rcsd7\" (UID: \"34ecc5fd-acf1-4748-af5d-d9af38bff6d3\") " pod="openshift-apiserver/apiserver-76f77b778f-rcsd7" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.111531 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/191743db-39d2-4f93-add6-471de6ccdede-encryption-config\") pod \"apiserver-7bbb656c7d-zv6wk\" (UID: \"191743db-39d2-4f93-add6-471de6ccdede\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zv6wk" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.111557 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e204bf3-844c-460e-9f4a-7d78025a3b09-config\") pod \"openshift-apiserver-operator-796bbdcf4f-kp5gm\" (UID: \"7e204bf3-844c-460e-9f4a-7d78025a3b09\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kp5gm" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.111580 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-vdxqp\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.111620 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/bfdd16e4-1cb8-4178-af49-4db764abc507-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-85llw\" (UID: \"bfdd16e4-1cb8-4178-af49-4db764abc507\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-85llw" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.111721 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2baa9d1-df52-44c6-b62f-f39c16b70e34-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-jq22x\" (UID: \"f2baa9d1-df52-44c6-b62f-f39c16b70e34\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jq22x" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.111757 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98bj5\" (UniqueName: \"kubernetes.io/projected/2f1d8a27-12bd-4ac8-b23f-51f9f10fbcbf-kube-api-access-98bj5\") pod \"catalog-operator-68c6474976-qz5br\" (UID: \"2f1d8a27-12bd-4ac8-b23f-51f9f10fbcbf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qz5br" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.111815 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/191743db-39d2-4f93-add6-471de6ccdede-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-zv6wk\" (UID: \"191743db-39d2-4f93-add6-471de6ccdede\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zv6wk" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.111857 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xstcq\" (UniqueName: \"kubernetes.io/projected/b22a0046-b9c4-4b6e-ba05-1c453c998ff8-kube-api-access-xstcq\") pod \"openshift-config-operator-7777fb866f-tqr7m\" (UID: \"b22a0046-b9c4-4b6e-ba05-1c453c998ff8\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-tqr7m" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.111882 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4bb8a0e-4090-4de8-b7b9-5b6fdc899cd0-config\") pod \"kube-apiserver-operator-766d6c64bb-4bfz8\" (UID: \"e4bb8a0e-4090-4de8-b7b9-5b6fdc899cd0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4bfz8" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.111915 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/f42f4e15-4186-4a35-aed6-122f4518d222-signing-cabundle\") pod \"service-ca-9c57cc56f-5btb4\" (UID: \"f42f4e15-4186-4a35-aed6-122f4518d222\") " pod="openshift-service-ca/service-ca-9c57cc56f-5btb4" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.111955 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/938dcbc0-ed30-4200-9f6f-cbc1364cce88-serving-cert\") pod \"console-operator-58897d9998-svdzq\" (UID: \"938dcbc0-ed30-4200-9f6f-cbc1364cce88\") " pod="openshift-console-operator/console-operator-58897d9998-svdzq" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.111990 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1c7126e3-85fe-4158-b92c-df2eb2717217-serving-cert\") pod \"controller-manager-879f6c89f-ddl6h\" (UID: \"1c7126e3-85fe-4158-b92c-df2eb2717217\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ddl6h" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.112009 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1c7126e3-85fe-4158-b92c-df2eb2717217-client-ca\") pod \"controller-manager-879f6c89f-ddl6h\" (UID: \"1c7126e3-85fe-4158-b92c-df2eb2717217\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ddl6h" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.112052 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqwlf\" (UniqueName: \"kubernetes.io/projected/9f245325-f6b0-4ffa-b423-63a47df33136-kube-api-access-mqwlf\") pod \"machine-approver-56656f9798-6jsdb\" (UID: \"9f245325-f6b0-4ffa-b423-63a47df33136\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6jsdb" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.112070 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/34ecc5fd-acf1-4748-af5d-d9af38bff6d3-node-pullsecrets\") pod \"apiserver-76f77b778f-rcsd7\" (UID: \"34ecc5fd-acf1-4748-af5d-d9af38bff6d3\") " pod="openshift-apiserver/apiserver-76f77b778f-rcsd7" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.112093 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b6794cf7-9161-499d-ba18-74b95caaf3ae-console-oauth-config\") pod \"console-f9d7485db-2qb64\" (UID: \"b6794cf7-9161-499d-ba18-74b95caaf3ae\") " pod="openshift-console/console-f9d7485db-2qb64" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.112115 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fecaea20-4e1a-4977-92e0-d0bd7c3037a2-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-rhzbg\" (UID: \"fecaea20-4e1a-4977-92e0-d0bd7c3037a2\") " pod="openshift-marketplace/marketplace-operator-79b997595-rhzbg" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.112137 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/fecaea20-4e1a-4977-92e0-d0bd7c3037a2-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-rhzbg\" (UID: \"fecaea20-4e1a-4977-92e0-d0bd7c3037a2\") " pod="openshift-marketplace/marketplace-operator-79b997595-rhzbg" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.112187 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3a945cd2-2261-435a-8d3f-bacd5b808d26-srv-cert\") pod \"olm-operator-6b444d44fb-9krqq\" (UID: \"3a945cd2-2261-435a-8d3f-bacd5b808d26\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9krqq" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.112229 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a98eb357-7807-407c-a1e0-9d9b2b94e1fa-serving-cert\") pod \"route-controller-manager-6576b87f9c-z7df5\" (UID: \"a98eb357-7807-407c-a1e0-9d9b2b94e1fa\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-z7df5" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.112381 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e4bb8a0e-4090-4de8-b7b9-5b6fdc899cd0-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-4bfz8\" (UID: \"e4bb8a0e-4090-4de8-b7b9-5b6fdc899cd0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4bfz8" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.112472 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sffxt\" (UniqueName: \"kubernetes.io/projected/f42f4e15-4186-4a35-aed6-122f4518d222-kube-api-access-sffxt\") pod \"service-ca-9c57cc56f-5btb4\" (UID: \"f42f4e15-4186-4a35-aed6-122f4518d222\") " pod="openshift-service-ca/service-ca-9c57cc56f-5btb4" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.112497 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74f1d4e8-2506-43c8-bcbb-cb8bf28cfb35-config\") pod \"kube-controller-manager-operator-78b949d7b-xj4jw\" (UID: \"74f1d4e8-2506-43c8-bcbb-cb8bf28cfb35\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xj4jw" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.112521 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2f1d8a27-12bd-4ac8-b23f-51f9f10fbcbf-profile-collector-cert\") pod \"catalog-operator-68c6474976-qz5br\" (UID: \"2f1d8a27-12bd-4ac8-b23f-51f9f10fbcbf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qz5br" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.112561 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-vdxqp\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.112624 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzcrd\" (UniqueName: \"kubernetes.io/projected/e69469d9-e3b1-4afd-84de-9b7dafa8e9e4-kube-api-access-lzcrd\") pod \"multus-admission-controller-857f4d67dd-2t46v\" (UID: \"e69469d9-e3b1-4afd-84de-9b7dafa8e9e4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-2t46v" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.112658 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-vdxqp\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.112688 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74f1d4e8-2506-43c8-bcbb-cb8bf28cfb35-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-xj4jw\" (UID: \"74f1d4e8-2506-43c8-bcbb-cb8bf28cfb35\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xj4jw" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.112740 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6743e3ae-38d5-4ac1-8eea-7f0fbbe282da-apiservice-cert\") pod \"packageserver-d55dfcdfc-g8b88\" (UID: \"6743e3ae-38d5-4ac1-8eea-7f0fbbe282da\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g8b88" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.112764 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1c7126e3-85fe-4158-b92c-df2eb2717217-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-ddl6h\" (UID: \"1c7126e3-85fe-4158-b92c-df2eb2717217\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ddl6h" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.112827 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e204bf3-844c-460e-9f4a-7d78025a3b09-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-kp5gm\" (UID: \"7e204bf3-844c-460e-9f4a-7d78025a3b09\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kp5gm" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.112863 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/74f1d4e8-2506-43c8-bcbb-cb8bf28cfb35-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-xj4jw\" (UID: \"74f1d4e8-2506-43c8-bcbb-cb8bf28cfb35\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xj4jw" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.112925 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b6794cf7-9161-499d-ba18-74b95caaf3ae-trusted-ca-bundle\") pod \"console-f9d7485db-2qb64\" (UID: \"b6794cf7-9161-499d-ba18-74b95caaf3ae\") " pod="openshift-console/console-f9d7485db-2qb64" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.112975 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9f2aa3db-bf5b-47b3-bcec-5132916f0f7d-etcd-client\") pod \"etcd-operator-b45778765-hkpdn\" (UID: \"9f2aa3db-bf5b-47b3-bcec-5132916f0f7d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hkpdn" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.113009 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/cddf4720-7bab-43e0-989e-6cff2942a5fe-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-qd7d2\" (UID: \"cddf4720-7bab-43e0-989e-6cff2942a5fe\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qd7d2" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.113049 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvdj4\" (UniqueName: \"kubernetes.io/projected/a98eb357-7807-407c-a1e0-9d9b2b94e1fa-kube-api-access-kvdj4\") pod \"route-controller-manager-6576b87f9c-z7df5\" (UID: \"a98eb357-7807-407c-a1e0-9d9b2b94e1fa\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-z7df5" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.113159 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/9f2aa3db-bf5b-47b3-bcec-5132916f0f7d-etcd-ca\") pod \"etcd-operator-b45778765-hkpdn\" (UID: \"9f2aa3db-bf5b-47b3-bcec-5132916f0f7d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hkpdn" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.113217 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/9f245325-f6b0-4ffa-b423-63a47df33136-machine-approver-tls\") pod \"machine-approver-56656f9798-6jsdb\" (UID: \"9f245325-f6b0-4ffa-b423-63a47df33136\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6jsdb" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.113321 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/191743db-39d2-4f93-add6-471de6ccdede-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-zv6wk\" (UID: \"191743db-39d2-4f93-add6-471de6ccdede\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zv6wk" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.113350 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-vdxqp\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.113378 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/34ecc5fd-acf1-4748-af5d-d9af38bff6d3-serving-cert\") pod \"apiserver-76f77b778f-rcsd7\" (UID: \"34ecc5fd-acf1-4748-af5d-d9af38bff6d3\") " pod="openshift-apiserver/apiserver-76f77b778f-rcsd7" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.113447 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6743e3ae-38d5-4ac1-8eea-7f0fbbe282da-webhook-cert\") pod \"packageserver-d55dfcdfc-g8b88\" (UID: \"6743e3ae-38d5-4ac1-8eea-7f0fbbe282da\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g8b88" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.113502 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6t28\" (UniqueName: \"kubernetes.io/projected/938dcbc0-ed30-4200-9f6f-cbc1364cce88-kube-api-access-v6t28\") pod \"console-operator-58897d9998-svdzq\" (UID: \"938dcbc0-ed30-4200-9f6f-cbc1364cce88\") " pod="openshift-console-operator/console-operator-58897d9998-svdzq" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.113557 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vv4zx\" (UniqueName: \"kubernetes.io/projected/9f2aa3db-bf5b-47b3-bcec-5132916f0f7d-kube-api-access-vv4zx\") pod \"etcd-operator-b45778765-hkpdn\" (UID: \"9f2aa3db-bf5b-47b3-bcec-5132916f0f7d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hkpdn" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.113593 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f245325-f6b0-4ffa-b423-63a47df33136-config\") pod \"machine-approver-56656f9798-6jsdb\" (UID: \"9f245325-f6b0-4ffa-b423-63a47df33136\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6jsdb" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.113640 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-vdxqp\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.113665 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-audit-dir\") pod \"oauth-openshift-558db77b4-vdxqp\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.113691 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7pq8\" (UniqueName: \"kubernetes.io/projected/36a3c13c-2670-4b33-a9fe-9071b35365fe-kube-api-access-r7pq8\") pod \"router-default-5444994796-jrhwp\" (UID: \"36a3c13c-2670-4b33-a9fe-9071b35365fe\") " pod="openshift-ingress/router-default-5444994796-jrhwp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.113724 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mctqc\" (UniqueName: \"kubernetes.io/projected/50764326-a80d-4b7a-bfd3-2242f4d0e228-kube-api-access-mctqc\") pod \"authentication-operator-69f744f599-fnhgz\" (UID: \"50764326-a80d-4b7a-bfd3-2242f4d0e228\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fnhgz" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.113752 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3a945cd2-2261-435a-8d3f-bacd5b808d26-profile-collector-cert\") pod \"olm-operator-6b444d44fb-9krqq\" (UID: \"3a945cd2-2261-435a-8d3f-bacd5b808d26\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9krqq" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.113984 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e204bf3-844c-460e-9f4a-7d78025a3b09-config\") pod \"openshift-apiserver-operator-796bbdcf4f-kp5gm\" (UID: \"7e204bf3-844c-460e-9f4a-7d78025a3b09\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kp5gm" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.118314 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f2aa3db-bf5b-47b3-bcec-5132916f0f7d-config\") pod \"etcd-operator-b45778765-hkpdn\" (UID: \"9f2aa3db-bf5b-47b3-bcec-5132916f0f7d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hkpdn" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.118371 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b6794cf7-9161-499d-ba18-74b95caaf3ae-service-ca\") pod \"console-f9d7485db-2qb64\" (UID: \"b6794cf7-9161-499d-ba18-74b95caaf3ae\") " pod="openshift-console/console-f9d7485db-2qb64" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.119734 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c7126e3-85fe-4158-b92c-df2eb2717217-config\") pod \"controller-manager-879f6c89f-ddl6h\" (UID: \"1c7126e3-85fe-4158-b92c-df2eb2717217\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ddl6h" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.119959 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a98eb357-7807-407c-a1e0-9d9b2b94e1fa-config\") pod \"route-controller-manager-6576b87f9c-z7df5\" (UID: \"a98eb357-7807-407c-a1e0-9d9b2b94e1fa\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-z7df5" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.120700 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-vdxqp\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.120722 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/938dcbc0-ed30-4200-9f6f-cbc1364cce88-trusted-ca\") pod \"console-operator-58897d9998-svdzq\" (UID: \"938dcbc0-ed30-4200-9f6f-cbc1364cce88\") " pod="openshift-console-operator/console-operator-58897d9998-svdzq" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.120799 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tbbpk\" (UniqueName: \"kubernetes.io/projected/7e204bf3-844c-460e-9f4a-7d78025a3b09-kube-api-access-tbbpk\") pod \"openshift-apiserver-operator-796bbdcf4f-kp5gm\" (UID: \"7e204bf3-844c-460e-9f4a-7d78025a3b09\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kp5gm" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.120831 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/50764326-a80d-4b7a-bfd3-2242f4d0e228-service-ca-bundle\") pod \"authentication-operator-69f744f599-fnhgz\" (UID: \"50764326-a80d-4b7a-bfd3-2242f4d0e228\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fnhgz" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.121260 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/b22a0046-b9c4-4b6e-ba05-1c453c998ff8-available-featuregates\") pod \"openshift-config-operator-7777fb866f-tqr7m\" (UID: \"b22a0046-b9c4-4b6e-ba05-1c453c998ff8\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-tqr7m" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.121301 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/34ecc5fd-acf1-4748-af5d-d9af38bff6d3-etcd-client\") pod \"apiserver-76f77b778f-rcsd7\" (UID: \"34ecc5fd-acf1-4748-af5d-d9af38bff6d3\") " pod="openshift-apiserver/apiserver-76f77b778f-rcsd7" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.121405 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/50764326-a80d-4b7a-bfd3-2242f4d0e228-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-fnhgz\" (UID: \"50764326-a80d-4b7a-bfd3-2242f4d0e228\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fnhgz" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.122034 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/50764326-a80d-4b7a-bfd3-2242f4d0e228-service-ca-bundle\") pod \"authentication-operator-69f744f599-fnhgz\" (UID: \"50764326-a80d-4b7a-bfd3-2242f4d0e228\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fnhgz" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.122251 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b6794cf7-9161-499d-ba18-74b95caaf3ae-oauth-serving-cert\") pod \"console-f9d7485db-2qb64\" (UID: \"b6794cf7-9161-499d-ba18-74b95caaf3ae\") " pod="openshift-console/console-f9d7485db-2qb64" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.122895 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/50764326-a80d-4b7a-bfd3-2242f4d0e228-config\") pod \"authentication-operator-69f744f599-fnhgz\" (UID: \"50764326-a80d-4b7a-bfd3-2242f4d0e228\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fnhgz" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.123350 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/b22a0046-b9c4-4b6e-ba05-1c453c998ff8-available-featuregates\") pod \"openshift-config-operator-7777fb866f-tqr7m\" (UID: \"b22a0046-b9c4-4b6e-ba05-1c453c998ff8\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-tqr7m" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.123338 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/34ecc5fd-acf1-4748-af5d-d9af38bff6d3-encryption-config\") pod \"apiserver-76f77b778f-rcsd7\" (UID: \"34ecc5fd-acf1-4748-af5d-d9af38bff6d3\") " pod="openshift-apiserver/apiserver-76f77b778f-rcsd7" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.123406 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/938dcbc0-ed30-4200-9f6f-cbc1364cce88-config\") pod \"console-operator-58897d9998-svdzq\" (UID: \"938dcbc0-ed30-4200-9f6f-cbc1364cce88\") " pod="openshift-console-operator/console-operator-58897d9998-svdzq" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.123437 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/34ecc5fd-acf1-4748-af5d-d9af38bff6d3-audit\") pod \"apiserver-76f77b778f-rcsd7\" (UID: \"34ecc5fd-acf1-4748-af5d-d9af38bff6d3\") " pod="openshift-apiserver/apiserver-76f77b778f-rcsd7" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.123464 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/191743db-39d2-4f93-add6-471de6ccdede-serving-cert\") pod \"apiserver-7bbb656c7d-zv6wk\" (UID: \"191743db-39d2-4f93-add6-471de6ccdede\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zv6wk" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.123498 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gdst\" (UniqueName: \"kubernetes.io/projected/f2baa9d1-df52-44c6-b62f-f39c16b70e34-kube-api-access-8gdst\") pod \"kube-storage-version-migrator-operator-b67b599dd-jq22x\" (UID: \"f2baa9d1-df52-44c6-b62f-f39c16b70e34\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jq22x" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.123536 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xb69\" (UniqueName: \"kubernetes.io/projected/ee4ae924-84bb-4b5c-9b6e-f03d6cba2ad3-kube-api-access-8xb69\") pod \"dns-operator-744455d44c-bpxtd\" (UID: \"ee4ae924-84bb-4b5c-9b6e-f03d6cba2ad3\") " pod="openshift-dns-operator/dns-operator-744455d44c-bpxtd" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.123589 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/34ecc5fd-acf1-4748-af5d-d9af38bff6d3-trusted-ca-bundle\") pod \"apiserver-76f77b778f-rcsd7\" (UID: \"34ecc5fd-acf1-4748-af5d-d9af38bff6d3\") " pod="openshift-apiserver/apiserver-76f77b778f-rcsd7" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.123592 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.123614 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-vdxqp\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.123641 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e4bb8a0e-4090-4de8-b7b9-5b6fdc899cd0-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-4bfz8\" (UID: \"e4bb8a0e-4090-4de8-b7b9-5b6fdc899cd0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4bfz8" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.123664 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/36a3c13c-2670-4b33-a9fe-9071b35365fe-stats-auth\") pod \"router-default-5444994796-jrhwp\" (UID: \"36a3c13c-2670-4b33-a9fe-9071b35365fe\") " pod="openshift-ingress/router-default-5444994796-jrhwp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.123825 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-vdxqp\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.123900 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9f245325-f6b0-4ffa-b423-63a47df33136-auth-proxy-config\") pod \"machine-approver-56656f9798-6jsdb\" (UID: \"9f245325-f6b0-4ffa-b423-63a47df33136\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6jsdb" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.123997 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlnqf\" (UniqueName: \"kubernetes.io/projected/191743db-39d2-4f93-add6-471de6ccdede-kube-api-access-rlnqf\") pod \"apiserver-7bbb656c7d-zv6wk\" (UID: \"191743db-39d2-4f93-add6-471de6ccdede\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zv6wk" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.124041 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b6794cf7-9161-499d-ba18-74b95caaf3ae-console-serving-cert\") pod \"console-f9d7485db-2qb64\" (UID: \"b6794cf7-9161-499d-ba18-74b95caaf3ae\") " pod="openshift-console/console-f9d7485db-2qb64" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.124076 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.124081 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/36a3c13c-2670-4b33-a9fe-9071b35365fe-service-ca-bundle\") pod \"router-default-5444994796-jrhwp\" (UID: \"36a3c13c-2670-4b33-a9fe-9071b35365fe\") " pod="openshift-ingress/router-default-5444994796-jrhwp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.124313 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7psw\" (UniqueName: \"kubernetes.io/projected/3a945cd2-2261-435a-8d3f-bacd5b808d26-kube-api-access-l7psw\") pod \"olm-operator-6b444d44fb-9krqq\" (UID: \"3a945cd2-2261-435a-8d3f-bacd5b808d26\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9krqq" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.124408 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-vdxqp\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.124454 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34ecc5fd-acf1-4748-af5d-d9af38bff6d3-config\") pod \"apiserver-76f77b778f-rcsd7\" (UID: \"34ecc5fd-acf1-4748-af5d-d9af38bff6d3\") " pod="openshift-apiserver/apiserver-76f77b778f-rcsd7" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.124535 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfk8z\" (UniqueName: \"kubernetes.io/projected/34ecc5fd-acf1-4748-af5d-d9af38bff6d3-kube-api-access-sfk8z\") pod \"apiserver-76f77b778f-rcsd7\" (UID: \"34ecc5fd-acf1-4748-af5d-d9af38bff6d3\") " pod="openshift-apiserver/apiserver-76f77b778f-rcsd7" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.124583 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svwvw\" (UniqueName: \"kubernetes.io/projected/96a1e39b-0018-41b8-8d7c-9fa7697858e9-kube-api-access-svwvw\") pod \"collect-profiles-29406510-f2lz7\" (UID: \"96a1e39b-0018-41b8-8d7c-9fa7697858e9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406510-f2lz7" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.124640 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/19582088-032e-4b42-bc19-5080f7c97e13-metrics-tls\") pod \"ingress-operator-5b745b69d9-9j7n6\" (UID: \"19582088-032e-4b42-bc19-5080f7c97e13\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9j7n6" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.124745 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-audit-policies\") pod \"oauth-openshift-558db77b4-vdxqp\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.124802 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/34ecc5fd-acf1-4748-af5d-d9af38bff6d3-audit-dir\") pod \"apiserver-76f77b778f-rcsd7\" (UID: \"34ecc5fd-acf1-4748-af5d-d9af38bff6d3\") " pod="openshift-apiserver/apiserver-76f77b778f-rcsd7" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.124833 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/191743db-39d2-4f93-add6-471de6ccdede-etcd-client\") pod \"apiserver-7bbb656c7d-zv6wk\" (UID: \"191743db-39d2-4f93-add6-471de6ccdede\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zv6wk" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.124887 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/191743db-39d2-4f93-add6-471de6ccdede-audit-dir\") pod \"apiserver-7bbb656c7d-zv6wk\" (UID: \"191743db-39d2-4f93-add6-471de6ccdede\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zv6wk" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.124946 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/938dcbc0-ed30-4200-9f6f-cbc1364cce88-config\") pod \"console-operator-58897d9998-svdzq\" (UID: \"938dcbc0-ed30-4200-9f6f-cbc1364cce88\") " pod="openshift-console-operator/console-operator-58897d9998-svdzq" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.125361 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/191743db-39d2-4f93-add6-471de6ccdede-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-zv6wk\" (UID: \"191743db-39d2-4f93-add6-471de6ccdede\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zv6wk" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.124463 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/34ecc5fd-acf1-4748-af5d-d9af38bff6d3-image-import-ca\") pod \"apiserver-76f77b778f-rcsd7\" (UID: \"34ecc5fd-acf1-4748-af5d-d9af38bff6d3\") " pod="openshift-apiserver/apiserver-76f77b778f-rcsd7" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.125531 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/191743db-39d2-4f93-add6-471de6ccdede-audit-dir\") pod \"apiserver-7bbb656c7d-zv6wk\" (UID: \"191743db-39d2-4f93-add6-471de6ccdede\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zv6wk" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.126211 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/34ecc5fd-acf1-4748-af5d-d9af38bff6d3-audit\") pod \"apiserver-76f77b778f-rcsd7\" (UID: \"34ecc5fd-acf1-4748-af5d-d9af38bff6d3\") " pod="openshift-apiserver/apiserver-76f77b778f-rcsd7" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.126524 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f2aa3db-bf5b-47b3-bcec-5132916f0f7d-serving-cert\") pod \"etcd-operator-b45778765-hkpdn\" (UID: \"9f2aa3db-bf5b-47b3-bcec-5132916f0f7d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hkpdn" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.127405 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/34ecc5fd-acf1-4748-af5d-d9af38bff6d3-trusted-ca-bundle\") pod \"apiserver-76f77b778f-rcsd7\" (UID: \"34ecc5fd-acf1-4748-af5d-d9af38bff6d3\") " pod="openshift-apiserver/apiserver-76f77b778f-rcsd7" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.128937 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c7126e3-85fe-4158-b92c-df2eb2717217-config\") pod \"controller-manager-879f6c89f-ddl6h\" (UID: \"1c7126e3-85fe-4158-b92c-df2eb2717217\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ddl6h" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.129182 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qz5br"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.129223 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-2bbp8"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.129244 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-svdzq"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.129547 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/191743db-39d2-4f93-add6-471de6ccdede-serving-cert\") pod \"apiserver-7bbb656c7d-zv6wk\" (UID: \"191743db-39d2-4f93-add6-471de6ccdede\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zv6wk" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.130527 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/938dcbc0-ed30-4200-9f6f-cbc1364cce88-trusted-ca\") pod \"console-operator-58897d9998-svdzq\" (UID: \"938dcbc0-ed30-4200-9f6f-cbc1364cce88\") " pod="openshift-console-operator/console-operator-58897d9998-svdzq" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.131145 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-z7df5"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.132925 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-vdxqp\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.142044 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.142085 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a98eb357-7807-407c-a1e0-9d9b2b94e1fa-config\") pod \"route-controller-manager-6576b87f9c-z7df5\" (UID: \"a98eb357-7807-407c-a1e0-9d9b2b94e1fa\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-z7df5" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.142654 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-6gm29"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.143751 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jq22x"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.143946 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/34ecc5fd-acf1-4748-af5d-d9af38bff6d3-etcd-serving-ca\") pod \"apiserver-76f77b778f-rcsd7\" (UID: \"34ecc5fd-acf1-4748-af5d-d9af38bff6d3\") " pod="openshift-apiserver/apiserver-76f77b778f-rcsd7" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.144617 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/191743db-39d2-4f93-add6-471de6ccdede-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-zv6wk\" (UID: \"191743db-39d2-4f93-add6-471de6ccdede\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zv6wk" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.145012 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-vdxqp\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.145560 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34ecc5fd-acf1-4748-af5d-d9af38bff6d3-config\") pod \"apiserver-76f77b778f-rcsd7\" (UID: \"34ecc5fd-acf1-4748-af5d-d9af38bff6d3\") " pod="openshift-apiserver/apiserver-76f77b778f-rcsd7" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.145604 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/34ecc5fd-acf1-4748-af5d-d9af38bff6d3-audit-dir\") pod \"apiserver-76f77b778f-rcsd7\" (UID: \"34ecc5fd-acf1-4748-af5d-d9af38bff6d3\") " pod="openshift-apiserver/apiserver-76f77b778f-rcsd7" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.146092 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9f245325-f6b0-4ffa-b423-63a47df33136-auth-proxy-config\") pod \"machine-approver-56656f9798-6jsdb\" (UID: \"9f245325-f6b0-4ffa-b423-63a47df33136\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6jsdb" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.147016 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4bfz8"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.147059 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-t4f4r"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.147073 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xj4jw"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.147506 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b6794cf7-9161-499d-ba18-74b95caaf3ae-trusted-ca-bundle\") pod \"console-f9d7485db-2qb64\" (UID: \"b6794cf7-9161-499d-ba18-74b95caaf3ae\") " pod="openshift-console/console-f9d7485db-2qb64" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.147888 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-vdxqp\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.149999 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9f2aa3db-bf5b-47b3-bcec-5132916f0f7d-etcd-client\") pod \"etcd-operator-b45778765-hkpdn\" (UID: \"9f2aa3db-bf5b-47b3-bcec-5132916f0f7d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hkpdn" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.150022 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1c7126e3-85fe-4158-b92c-df2eb2717217-serving-cert\") pod \"controller-manager-879f6c89f-ddl6h\" (UID: \"1c7126e3-85fe-4158-b92c-df2eb2717217\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ddl6h" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.150241 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/191743db-39d2-4f93-add6-471de6ccdede-encryption-config\") pod \"apiserver-7bbb656c7d-zv6wk\" (UID: \"191743db-39d2-4f93-add6-471de6ccdede\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zv6wk" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.150588 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a98eb357-7807-407c-a1e0-9d9b2b94e1fa-client-ca\") pod \"route-controller-manager-6576b87f9c-z7df5\" (UID: \"a98eb357-7807-407c-a1e0-9d9b2b94e1fa\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-z7df5" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.150823 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/9f2aa3db-bf5b-47b3-bcec-5132916f0f7d-etcd-service-ca\") pod \"etcd-operator-b45778765-hkpdn\" (UID: \"9f2aa3db-bf5b-47b3-bcec-5132916f0f7d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hkpdn" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.150880 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/34ecc5fd-acf1-4748-af5d-d9af38bff6d3-node-pullsecrets\") pod \"apiserver-76f77b778f-rcsd7\" (UID: \"34ecc5fd-acf1-4748-af5d-d9af38bff6d3\") " pod="openshift-apiserver/apiserver-76f77b778f-rcsd7" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.151369 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b22a0046-b9c4-4b6e-ba05-1c453c998ff8-serving-cert\") pod \"openshift-config-operator-7777fb866f-tqr7m\" (UID: \"b22a0046-b9c4-4b6e-ba05-1c453c998ff8\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-tqr7m" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.151621 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1c7126e3-85fe-4158-b92c-df2eb2717217-client-ca\") pod \"controller-manager-879f6c89f-ddl6h\" (UID: \"1c7126e3-85fe-4158-b92c-df2eb2717217\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ddl6h" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.152014 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-jjnx9"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.152366 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/34ecc5fd-acf1-4748-af5d-d9af38bff6d3-serving-cert\") pod \"apiserver-76f77b778f-rcsd7\" (UID: \"34ecc5fd-acf1-4748-af5d-d9af38bff6d3\") " pod="openshift-apiserver/apiserver-76f77b778f-rcsd7" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.152692 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-bbm46"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.152714 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-audit-dir\") pod \"oauth-openshift-558db77b4-vdxqp\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.152717 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-2qb64"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.152813 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-jjnx9" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.153132 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f245325-f6b0-4ffa-b423-63a47df33136-config\") pod \"machine-approver-56656f9798-6jsdb\" (UID: \"9f245325-f6b0-4ffa-b423-63a47df33136\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6jsdb" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.153564 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-vdxqp\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.153606 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e204bf3-844c-460e-9f4a-7d78025a3b09-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-kp5gm\" (UID: \"7e204bf3-844c-460e-9f4a-7d78025a3b09\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kp5gm" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.154772 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42edbb7d-0d06-4a5f-b481-7345a1b67227-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-brs97\" (UID: \"42edbb7d-0d06-4a5f-b481-7345a1b67227\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-brs97" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.155116 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/50764326-a80d-4b7a-bfd3-2242f4d0e228-serving-cert\") pod \"authentication-operator-69f744f599-fnhgz\" (UID: \"50764326-a80d-4b7a-bfd3-2242f4d0e228\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fnhgz" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.155421 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.155734 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1c7126e3-85fe-4158-b92c-df2eb2717217-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-ddl6h\" (UID: \"1c7126e3-85fe-4158-b92c-df2eb2717217\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ddl6h" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.155884 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e73c6003-61f1-49b4-9275-14a55a35e186-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-6gm29\" (UID: \"e73c6003-61f1-49b4-9275-14a55a35e186\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6gm29" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.156146 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b6794cf7-9161-499d-ba18-74b95caaf3ae-console-oauth-config\") pod \"console-f9d7485db-2qb64\" (UID: \"b6794cf7-9161-499d-ba18-74b95caaf3ae\") " pod="openshift-console/console-f9d7485db-2qb64" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.156718 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ee4ae924-84bb-4b5c-9b6e-f03d6cba2ad3-metrics-tls\") pod \"dns-operator-744455d44c-bpxtd\" (UID: \"ee4ae924-84bb-4b5c-9b6e-f03d6cba2ad3\") " pod="openshift-dns-operator/dns-operator-744455d44c-bpxtd" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.157372 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/938dcbc0-ed30-4200-9f6f-cbc1364cce88-serving-cert\") pod \"console-operator-58897d9998-svdzq\" (UID: \"938dcbc0-ed30-4200-9f6f-cbc1364cce88\") " pod="openshift-console-operator/console-operator-58897d9998-svdzq" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.157451 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-2wtrf"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.157816 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b6794cf7-9161-499d-ba18-74b95caaf3ae-console-serving-cert\") pod \"console-f9d7485db-2qb64\" (UID: \"b6794cf7-9161-499d-ba18-74b95caaf3ae\") " pod="openshift-console/console-f9d7485db-2qb64" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.158373 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-vdxqp\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.158856 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-2wtrf" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.158897 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/42edbb7d-0d06-4a5f-b481-7345a1b67227-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-brs97\" (UID: \"42edbb7d-0d06-4a5f-b481-7345a1b67227\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-brs97" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.159190 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4bb8a0e-4090-4de8-b7b9-5b6fdc899cd0-config\") pod \"kube-apiserver-operator-766d6c64bb-4bfz8\" (UID: \"e4bb8a0e-4090-4de8-b7b9-5b6fdc899cd0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4bfz8" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.159504 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/e86c829a-524d-4194-9ab2-717b21d7cd60-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-vk2xx\" (UID: \"e86c829a-524d-4194-9ab2-717b21d7cd60\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vk2xx" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.159516 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/191743db-39d2-4f93-add6-471de6ccdede-audit-policies\") pod \"apiserver-7bbb656c7d-zv6wk\" (UID: \"191743db-39d2-4f93-add6-471de6ccdede\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zv6wk" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.159974 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e4bb8a0e-4090-4de8-b7b9-5b6fdc899cd0-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-4bfz8\" (UID: \"e4bb8a0e-4090-4de8-b7b9-5b6fdc899cd0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4bfz8" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.159990 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-vbffw"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.160355 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a98eb357-7807-407c-a1e0-9d9b2b94e1fa-serving-cert\") pod \"route-controller-manager-6576b87f9c-z7df5\" (UID: \"a98eb357-7807-407c-a1e0-9d9b2b94e1fa\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-z7df5" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.160106 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b6794cf7-9161-499d-ba18-74b95caaf3ae-console-config\") pod \"console-f9d7485db-2qb64\" (UID: \"b6794cf7-9161-499d-ba18-74b95caaf3ae\") " pod="openshift-console/console-f9d7485db-2qb64" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.160890 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/191743db-39d2-4f93-add6-471de6ccdede-etcd-client\") pod \"apiserver-7bbb656c7d-zv6wk\" (UID: \"191743db-39d2-4f93-add6-471de6ccdede\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zv6wk" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.161704 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h4w54"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.161927 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-vbffw" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.163028 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-vdxqp\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.166247 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/9f245325-f6b0-4ffa-b423-63a47df33136-machine-approver-tls\") pod \"machine-approver-56656f9798-6jsdb\" (UID: \"9f245325-f6b0-4ffa-b423-63a47df33136\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6jsdb" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.166495 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-vdxqp\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.166561 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-5btb4"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.166635 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-vdxqp\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.168225 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-6rp4p"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.170125 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.170477 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-nmvwb"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.173519 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hs96x"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.174222 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9krqq"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.176726 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-vdxqp\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.177050 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-jjnx9"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.179005 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-vbffw"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.180165 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rhzbg"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.181423 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406510-f2lz7"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.183569 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qd7d2"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.184988 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g8b88"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.185934 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-vh8lz"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.187271 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-vh8lz"] Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.187289 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-vh8lz" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.190635 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.210835 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.215854 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-vdxqp\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.226300 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/f42f4e15-4186-4a35-aed6-122f4518d222-signing-cabundle\") pod \"service-ca-9c57cc56f-5btb4\" (UID: \"f42f4e15-4186-4a35-aed6-122f4518d222\") " pod="openshift-service-ca/service-ca-9c57cc56f-5btb4" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.226345 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/fecaea20-4e1a-4977-92e0-d0bd7c3037a2-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-rhzbg\" (UID: \"fecaea20-4e1a-4977-92e0-d0bd7c3037a2\") " pod="openshift-marketplace/marketplace-operator-79b997595-rhzbg" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.226446 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3a945cd2-2261-435a-8d3f-bacd5b808d26-srv-cert\") pod \"olm-operator-6b444d44fb-9krqq\" (UID: \"3a945cd2-2261-435a-8d3f-bacd5b808d26\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9krqq" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.226479 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fecaea20-4e1a-4977-92e0-d0bd7c3037a2-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-rhzbg\" (UID: \"fecaea20-4e1a-4977-92e0-d0bd7c3037a2\") " pod="openshift-marketplace/marketplace-operator-79b997595-rhzbg" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.226499 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sffxt\" (UniqueName: \"kubernetes.io/projected/f42f4e15-4186-4a35-aed6-122f4518d222-kube-api-access-sffxt\") pod \"service-ca-9c57cc56f-5btb4\" (UID: \"f42f4e15-4186-4a35-aed6-122f4518d222\") " pod="openshift-service-ca/service-ca-9c57cc56f-5btb4" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.226528 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2f1d8a27-12bd-4ac8-b23f-51f9f10fbcbf-profile-collector-cert\") pod \"catalog-operator-68c6474976-qz5br\" (UID: \"2f1d8a27-12bd-4ac8-b23f-51f9f10fbcbf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qz5br" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.226548 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzcrd\" (UniqueName: \"kubernetes.io/projected/e69469d9-e3b1-4afd-84de-9b7dafa8e9e4-kube-api-access-lzcrd\") pod \"multus-admission-controller-857f4d67dd-2t46v\" (UID: \"e69469d9-e3b1-4afd-84de-9b7dafa8e9e4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-2t46v" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.226575 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6743e3ae-38d5-4ac1-8eea-7f0fbbe282da-apiservice-cert\") pod \"packageserver-d55dfcdfc-g8b88\" (UID: \"6743e3ae-38d5-4ac1-8eea-7f0fbbe282da\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g8b88" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.226604 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/cddf4720-7bab-43e0-989e-6cff2942a5fe-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-qd7d2\" (UID: \"cddf4720-7bab-43e0-989e-6cff2942a5fe\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qd7d2" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.226650 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6743e3ae-38d5-4ac1-8eea-7f0fbbe282da-webhook-cert\") pod \"packageserver-d55dfcdfc-g8b88\" (UID: \"6743e3ae-38d5-4ac1-8eea-7f0fbbe282da\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g8b88" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.226691 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3a945cd2-2261-435a-8d3f-bacd5b808d26-profile-collector-cert\") pod \"olm-operator-6b444d44fb-9krqq\" (UID: \"3a945cd2-2261-435a-8d3f-bacd5b808d26\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9krqq" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.226766 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8gdst\" (UniqueName: \"kubernetes.io/projected/f2baa9d1-df52-44c6-b62f-f39c16b70e34-kube-api-access-8gdst\") pod \"kube-storage-version-migrator-operator-b67b599dd-jq22x\" (UID: \"f2baa9d1-df52-44c6-b62f-f39c16b70e34\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jq22x" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.226933 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7psw\" (UniqueName: \"kubernetes.io/projected/3a945cd2-2261-435a-8d3f-bacd5b808d26-kube-api-access-l7psw\") pod \"olm-operator-6b444d44fb-9krqq\" (UID: \"3a945cd2-2261-435a-8d3f-bacd5b808d26\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9krqq" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.226968 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svwvw\" (UniqueName: \"kubernetes.io/projected/96a1e39b-0018-41b8-8d7c-9fa7697858e9-kube-api-access-svwvw\") pod \"collect-profiles-29406510-f2lz7\" (UID: \"96a1e39b-0018-41b8-8d7c-9fa7697858e9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406510-f2lz7" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.227017 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/f42f4e15-4186-4a35-aed6-122f4518d222-signing-key\") pod \"service-ca-9c57cc56f-5btb4\" (UID: \"f42f4e15-4186-4a35-aed6-122f4518d222\") " pod="openshift-service-ca/service-ca-9c57cc56f-5btb4" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.227052 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/6743e3ae-38d5-4ac1-8eea-7f0fbbe282da-tmpfs\") pod \"packageserver-d55dfcdfc-g8b88\" (UID: \"6743e3ae-38d5-4ac1-8eea-7f0fbbe282da\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g8b88" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.227082 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2f1d8a27-12bd-4ac8-b23f-51f9f10fbcbf-srv-cert\") pod \"catalog-operator-68c6474976-qz5br\" (UID: \"2f1d8a27-12bd-4ac8-b23f-51f9f10fbcbf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qz5br" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.227154 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/96a1e39b-0018-41b8-8d7c-9fa7697858e9-config-volume\") pod \"collect-profiles-29406510-f2lz7\" (UID: \"96a1e39b-0018-41b8-8d7c-9fa7697858e9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406510-f2lz7" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.227228 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e69469d9-e3b1-4afd-84de-9b7dafa8e9e4-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-2t46v\" (UID: \"e69469d9-e3b1-4afd-84de-9b7dafa8e9e4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-2t46v" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.227254 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nstwg\" (UniqueName: \"kubernetes.io/projected/6743e3ae-38d5-4ac1-8eea-7f0fbbe282da-kube-api-access-nstwg\") pod \"packageserver-d55dfcdfc-g8b88\" (UID: \"6743e3ae-38d5-4ac1-8eea-7f0fbbe282da\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g8b88" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.227294 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzpbl\" (UniqueName: \"kubernetes.io/projected/fecaea20-4e1a-4977-92e0-d0bd7c3037a2-kube-api-access-fzpbl\") pod \"marketplace-operator-79b997595-rhzbg\" (UID: \"fecaea20-4e1a-4977-92e0-d0bd7c3037a2\") " pod="openshift-marketplace/marketplace-operator-79b997595-rhzbg" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.227319 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aab927aa-d4cb-4b96-a5a6-bb14df8d4af8-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-hs96x\" (UID: \"aab927aa-d4cb-4b96-a5a6-bb14df8d4af8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hs96x" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.227343 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/aab927aa-d4cb-4b96-a5a6-bb14df8d4af8-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-hs96x\" (UID: \"aab927aa-d4cb-4b96-a5a6-bb14df8d4af8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hs96x" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.227360 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f2baa9d1-df52-44c6-b62f-f39c16b70e34-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-jq22x\" (UID: \"f2baa9d1-df52-44c6-b62f-f39c16b70e34\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jq22x" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.227381 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvj24\" (UniqueName: \"kubernetes.io/projected/cddf4720-7bab-43e0-989e-6cff2942a5fe-kube-api-access-nvj24\") pod \"package-server-manager-789f6589d5-qd7d2\" (UID: \"cddf4720-7bab-43e0-989e-6cff2942a5fe\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qd7d2" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.227420 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/96a1e39b-0018-41b8-8d7c-9fa7697858e9-secret-volume\") pod \"collect-profiles-29406510-f2lz7\" (UID: \"96a1e39b-0018-41b8-8d7c-9fa7697858e9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406510-f2lz7" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.227460 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aab927aa-d4cb-4b96-a5a6-bb14df8d4af8-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-hs96x\" (UID: \"aab927aa-d4cb-4b96-a5a6-bb14df8d4af8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hs96x" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.227491 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2baa9d1-df52-44c6-b62f-f39c16b70e34-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-jq22x\" (UID: \"f2baa9d1-df52-44c6-b62f-f39c16b70e34\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jq22x" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.227517 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98bj5\" (UniqueName: \"kubernetes.io/projected/2f1d8a27-12bd-4ac8-b23f-51f9f10fbcbf-kube-api-access-98bj5\") pod \"catalog-operator-68c6474976-qz5br\" (UID: \"2f1d8a27-12bd-4ac8-b23f-51f9f10fbcbf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qz5br" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.227927 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/6743e3ae-38d5-4ac1-8eea-7f0fbbe282da-tmpfs\") pod \"packageserver-d55dfcdfc-g8b88\" (UID: \"6743e3ae-38d5-4ac1-8eea-7f0fbbe282da\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g8b88" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.231652 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.236731 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-audit-policies\") pod \"oauth-openshift-558db77b4-vdxqp\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.250763 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.261408 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/9f2aa3db-bf5b-47b3-bcec-5132916f0f7d-etcd-ca\") pod \"etcd-operator-b45778765-hkpdn\" (UID: \"9f2aa3db-bf5b-47b3-bcec-5132916f0f7d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hkpdn" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.270839 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.297680 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.311086 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.311404 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/bfdd16e4-1cb8-4178-af49-4db764abc507-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-85llw\" (UID: \"bfdd16e4-1cb8-4178-af49-4db764abc507\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-85llw" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.330123 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.351100 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.371356 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.391986 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.397464 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74f1d4e8-2506-43c8-bcbb-cb8bf28cfb35-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-xj4jw\" (UID: \"74f1d4e8-2506-43c8-bcbb-cb8bf28cfb35\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xj4jw" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.415160 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.423194 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74f1d4e8-2506-43c8-bcbb-cb8bf28cfb35-config\") pod \"kube-controller-manager-operator-78b949d7b-xj4jw\" (UID: \"74f1d4e8-2506-43c8-bcbb-cb8bf28cfb35\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xj4jw" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.430251 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.438993 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/36a3c13c-2670-4b33-a9fe-9071b35365fe-default-certificate\") pod \"router-default-5444994796-jrhwp\" (UID: \"36a3c13c-2670-4b33-a9fe-9071b35365fe\") " pod="openshift-ingress/router-default-5444994796-jrhwp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.450394 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.457316 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/36a3c13c-2670-4b33-a9fe-9071b35365fe-stats-auth\") pod \"router-default-5444994796-jrhwp\" (UID: \"36a3c13c-2670-4b33-a9fe-9071b35365fe\") " pod="openshift-ingress/router-default-5444994796-jrhwp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.470209 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.474161 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/36a3c13c-2670-4b33-a9fe-9071b35365fe-metrics-certs\") pod \"router-default-5444994796-jrhwp\" (UID: \"36a3c13c-2670-4b33-a9fe-9071b35365fe\") " pod="openshift-ingress/router-default-5444994796-jrhwp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.490923 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.510387 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.515220 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/36a3c13c-2670-4b33-a9fe-9071b35365fe-service-ca-bundle\") pod \"router-default-5444994796-jrhwp\" (UID: \"36a3c13c-2670-4b33-a9fe-9071b35365fe\") " pod="openshift-ingress/router-default-5444994796-jrhwp" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.529928 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.550203 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.561647 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e73c6003-61f1-49b4-9275-14a55a35e186-proxy-tls\") pod \"machine-config-controller-84d6567774-6gm29\" (UID: \"e73c6003-61f1-49b4-9275-14a55a35e186\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6gm29" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.569925 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.610632 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.620214 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/19582088-032e-4b42-bc19-5080f7c97e13-metrics-tls\") pod \"ingress-operator-5b745b69d9-9j7n6\" (UID: \"19582088-032e-4b42-bc19-5080f7c97e13\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9j7n6" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.631400 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.649525 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.676252 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.682605 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/19582088-032e-4b42-bc19-5080f7c97e13-trusted-ca\") pod \"ingress-operator-5b745b69d9-9j7n6\" (UID: \"19582088-032e-4b42-bc19-5080f7c97e13\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9j7n6" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.690841 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.709964 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.720967 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f2baa9d1-df52-44c6-b62f-f39c16b70e34-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-jq22x\" (UID: \"f2baa9d1-df52-44c6-b62f-f39c16b70e34\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jq22x" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.730701 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.749738 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.769819 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.778402 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2baa9d1-df52-44c6-b62f-f39c16b70e34-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-jq22x\" (UID: \"f2baa9d1-df52-44c6-b62f-f39c16b70e34\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jq22x" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.790373 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.810485 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.820930 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/96a1e39b-0018-41b8-8d7c-9fa7697858e9-secret-volume\") pod \"collect-profiles-29406510-f2lz7\" (UID: \"96a1e39b-0018-41b8-8d7c-9fa7697858e9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406510-f2lz7" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.820930 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2f1d8a27-12bd-4ac8-b23f-51f9f10fbcbf-profile-collector-cert\") pod \"catalog-operator-68c6474976-qz5br\" (UID: \"2f1d8a27-12bd-4ac8-b23f-51f9f10fbcbf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qz5br" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.821343 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3a945cd2-2261-435a-8d3f-bacd5b808d26-profile-collector-cert\") pod \"olm-operator-6b444d44fb-9krqq\" (UID: \"3a945cd2-2261-435a-8d3f-bacd5b808d26\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9krqq" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.830015 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.840310 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2f1d8a27-12bd-4ac8-b23f-51f9f10fbcbf-srv-cert\") pod \"catalog-operator-68c6474976-qz5br\" (UID: \"2f1d8a27-12bd-4ac8-b23f-51f9f10fbcbf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qz5br" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.849235 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.869829 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.879348 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aab927aa-d4cb-4b96-a5a6-bb14df8d4af8-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-hs96x\" (UID: \"aab927aa-d4cb-4b96-a5a6-bb14df8d4af8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hs96x" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.890307 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.909885 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.929904 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.963663 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6nf4\" (UniqueName: \"kubernetes.io/projected/daa3f105-2960-45a0-ab05-cb2cedea7e9f-kube-api-access-g6nf4\") pod \"machine-api-operator-5694c8668f-wznnv\" (UID: \"daa3f105-2960-45a0-ab05-cb2cedea7e9f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wznnv" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.970131 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 29 04:40:12 crc kubenswrapper[4799]: I1129 04:40:12.990286 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.008848 4799 request.go:700] Waited for 1.003706205s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/secrets?fieldSelector=metadata.name%3Dmachine-config-operator-dockercfg-98p87&limit=500&resourceVersion=0 Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.010641 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.031066 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.051144 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.061204 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aab927aa-d4cb-4b96-a5a6-bb14df8d4af8-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-hs96x\" (UID: \"aab927aa-d4cb-4b96-a5a6-bb14df8d4af8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hs96x" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.068255 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-wznnv" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.076985 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.090602 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.110237 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.129757 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.150036 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.170153 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.190807 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.210835 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 29 04:40:13 crc kubenswrapper[4799]: E1129 04:40:13.227311 4799 secret.go:188] Couldn't get secret openshift-service-ca/signing-key: failed to sync secret cache: timed out waiting for the condition Nov 29 04:40:13 crc kubenswrapper[4799]: E1129 04:40:13.227351 4799 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/package-server-manager-serving-cert: failed to sync secret cache: timed out waiting for the condition Nov 29 04:40:13 crc kubenswrapper[4799]: E1129 04:40:13.231356 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cddf4720-7bab-43e0-989e-6cff2942a5fe-package-server-manager-serving-cert podName:cddf4720-7bab-43e0-989e-6cff2942a5fe nodeName:}" failed. No retries permitted until 2025-11-29 04:40:13.731320098 +0000 UTC m=+89.374250498 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "package-server-manager-serving-cert" (UniqueName: "kubernetes.io/secret/cddf4720-7bab-43e0-989e-6cff2942a5fe-package-server-manager-serving-cert") pod "package-server-manager-789f6589d5-qd7d2" (UID: "cddf4720-7bab-43e0-989e-6cff2942a5fe") : failed to sync secret cache: timed out waiting for the condition Nov 29 04:40:13 crc kubenswrapper[4799]: E1129 04:40:13.231446 4799 secret.go:188] Couldn't get secret openshift-multus/multus-admission-controller-secret: failed to sync secret cache: timed out waiting for the condition Nov 29 04:40:13 crc kubenswrapper[4799]: E1129 04:40:13.231556 4799 configmap.go:193] Couldn't get configMap openshift-service-ca/signing-cabundle: failed to sync configmap cache: timed out waiting for the condition Nov 29 04:40:13 crc kubenswrapper[4799]: E1129 04:40:13.231608 4799 secret.go:188] Couldn't get secret openshift-marketplace/marketplace-operator-metrics: failed to sync secret cache: timed out waiting for the condition Nov 29 04:40:13 crc kubenswrapper[4799]: E1129 04:40:13.231654 4799 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/packageserver-service-cert: failed to sync secret cache: timed out waiting for the condition Nov 29 04:40:13 crc kubenswrapper[4799]: E1129 04:40:13.231707 4799 configmap.go:193] Couldn't get configMap openshift-operator-lifecycle-manager/collect-profiles-config: failed to sync configmap cache: timed out waiting for the condition Nov 29 04:40:13 crc kubenswrapper[4799]: E1129 04:40:13.231712 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f42f4e15-4186-4a35-aed6-122f4518d222-signing-key podName:f42f4e15-4186-4a35-aed6-122f4518d222 nodeName:}" failed. No retries permitted until 2025-11-29 04:40:13.731490512 +0000 UTC m=+89.374420912 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "signing-key" (UniqueName: "kubernetes.io/secret/f42f4e15-4186-4a35-aed6-122f4518d222-signing-key") pod "service-ca-9c57cc56f-5btb4" (UID: "f42f4e15-4186-4a35-aed6-122f4518d222") : failed to sync secret cache: timed out waiting for the condition Nov 29 04:40:13 crc kubenswrapper[4799]: E1129 04:40:13.227389 4799 configmap.go:193] Couldn't get configMap openshift-marketplace/marketplace-trusted-ca: failed to sync configmap cache: timed out waiting for the condition Nov 29 04:40:13 crc kubenswrapper[4799]: E1129 04:40:13.231751 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e69469d9-e3b1-4afd-84de-9b7dafa8e9e4-webhook-certs podName:e69469d9-e3b1-4afd-84de-9b7dafa8e9e4 nodeName:}" failed. No retries permitted until 2025-11-29 04:40:13.731730268 +0000 UTC m=+89.374660668 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/e69469d9-e3b1-4afd-84de-9b7dafa8e9e4-webhook-certs") pod "multus-admission-controller-857f4d67dd-2t46v" (UID: "e69469d9-e3b1-4afd-84de-9b7dafa8e9e4") : failed to sync secret cache: timed out waiting for the condition Nov 29 04:40:13 crc kubenswrapper[4799]: E1129 04:40:13.231821 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/f42f4e15-4186-4a35-aed6-122f4518d222-signing-cabundle podName:f42f4e15-4186-4a35-aed6-122f4518d222 nodeName:}" failed. No retries permitted until 2025-11-29 04:40:13.73178179 +0000 UTC m=+89.374712190 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "signing-cabundle" (UniqueName: "kubernetes.io/configmap/f42f4e15-4186-4a35-aed6-122f4518d222-signing-cabundle") pod "service-ca-9c57cc56f-5btb4" (UID: "f42f4e15-4186-4a35-aed6-122f4518d222") : failed to sync configmap cache: timed out waiting for the condition Nov 29 04:40:13 crc kubenswrapper[4799]: E1129 04:40:13.231842 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fecaea20-4e1a-4977-92e0-d0bd7c3037a2-marketplace-operator-metrics podName:fecaea20-4e1a-4977-92e0-d0bd7c3037a2 nodeName:}" failed. No retries permitted until 2025-11-29 04:40:13.731834321 +0000 UTC m=+89.374764721 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "marketplace-operator-metrics" (UniqueName: "kubernetes.io/secret/fecaea20-4e1a-4977-92e0-d0bd7c3037a2-marketplace-operator-metrics") pod "marketplace-operator-79b997595-rhzbg" (UID: "fecaea20-4e1a-4977-92e0-d0bd7c3037a2") : failed to sync secret cache: timed out waiting for the condition Nov 29 04:40:13 crc kubenswrapper[4799]: E1129 04:40:13.231864 4799 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/packageserver-service-cert: failed to sync secret cache: timed out waiting for the condition Nov 29 04:40:13 crc kubenswrapper[4799]: E1129 04:40:13.231892 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6743e3ae-38d5-4ac1-8eea-7f0fbbe282da-apiservice-cert podName:6743e3ae-38d5-4ac1-8eea-7f0fbbe282da nodeName:}" failed. No retries permitted until 2025-11-29 04:40:13.731876192 +0000 UTC m=+89.374806602 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "apiservice-cert" (UniqueName: "kubernetes.io/secret/6743e3ae-38d5-4ac1-8eea-7f0fbbe282da-apiservice-cert") pod "packageserver-d55dfcdfc-g8b88" (UID: "6743e3ae-38d5-4ac1-8eea-7f0fbbe282da") : failed to sync secret cache: timed out waiting for the condition Nov 29 04:40:13 crc kubenswrapper[4799]: E1129 04:40:13.231910 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/96a1e39b-0018-41b8-8d7c-9fa7697858e9-config-volume podName:96a1e39b-0018-41b8-8d7c-9fa7697858e9 nodeName:}" failed. No retries permitted until 2025-11-29 04:40:13.731900653 +0000 UTC m=+89.374831063 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/96a1e39b-0018-41b8-8d7c-9fa7697858e9-config-volume") pod "collect-profiles-29406510-f2lz7" (UID: "96a1e39b-0018-41b8-8d7c-9fa7697858e9") : failed to sync configmap cache: timed out waiting for the condition Nov 29 04:40:13 crc kubenswrapper[4799]: E1129 04:40:13.231985 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/fecaea20-4e1a-4977-92e0-d0bd7c3037a2-marketplace-trusted-ca podName:fecaea20-4e1a-4977-92e0-d0bd7c3037a2 nodeName:}" failed. No retries permitted until 2025-11-29 04:40:13.731968944 +0000 UTC m=+89.374899344 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "marketplace-trusted-ca" (UniqueName: "kubernetes.io/configmap/fecaea20-4e1a-4977-92e0-d0bd7c3037a2-marketplace-trusted-ca") pod "marketplace-operator-79b997595-rhzbg" (UID: "fecaea20-4e1a-4977-92e0-d0bd7c3037a2") : failed to sync configmap cache: timed out waiting for the condition Nov 29 04:40:13 crc kubenswrapper[4799]: E1129 04:40:13.232056 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6743e3ae-38d5-4ac1-8eea-7f0fbbe282da-webhook-cert podName:6743e3ae-38d5-4ac1-8eea-7f0fbbe282da nodeName:}" failed. No retries permitted until 2025-11-29 04:40:13.731993935 +0000 UTC m=+89.374924335 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-cert" (UniqueName: "kubernetes.io/secret/6743e3ae-38d5-4ac1-8eea-7f0fbbe282da-webhook-cert") pod "packageserver-d55dfcdfc-g8b88" (UID: "6743e3ae-38d5-4ac1-8eea-7f0fbbe282da") : failed to sync secret cache: timed out waiting for the condition Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.233090 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.238176 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3a945cd2-2261-435a-8d3f-bacd5b808d26-srv-cert\") pod \"olm-operator-6b444d44fb-9krqq\" (UID: \"3a945cd2-2261-435a-8d3f-bacd5b808d26\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9krqq" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.244660 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-wznnv"] Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.250064 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.272832 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.282593 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-wznnv" event={"ID":"daa3f105-2960-45a0-ab05-cb2cedea7e9f","Type":"ContainerStarted","Data":"df1c28dc9e6fb6973e23e2f7e1913003430caefefb8bd1aaa3eba0e789c6be99"} Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.290195 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.309803 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.331715 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.349615 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.377440 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.391143 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.410194 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.429813 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.449916 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.470136 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.490362 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.510500 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.530334 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.549988 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.570352 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.591146 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.618242 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.630021 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.650473 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.658815 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.658843 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.658848 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qx9wn" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.658963 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.670722 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.725413 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfljm\" (UniqueName: \"kubernetes.io/projected/e73c6003-61f1-49b4-9275-14a55a35e186-kube-api-access-tfljm\") pod \"machine-config-controller-84d6567774-6gm29\" (UID: \"e73c6003-61f1-49b4-9275-14a55a35e186\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6gm29" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.743995 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q67cq\" (UniqueName: \"kubernetes.io/projected/e86c829a-524d-4194-9ab2-717b21d7cd60-kube-api-access-q67cq\") pod \"cluster-samples-operator-665b6dd947-vk2xx\" (UID: \"e86c829a-524d-4194-9ab2-717b21d7cd60\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vk2xx" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.746554 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6gm29" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.749856 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/f42f4e15-4186-4a35-aed6-122f4518d222-signing-cabundle\") pod \"service-ca-9c57cc56f-5btb4\" (UID: \"f42f4e15-4186-4a35-aed6-122f4518d222\") " pod="openshift-service-ca/service-ca-9c57cc56f-5btb4" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.749893 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fecaea20-4e1a-4977-92e0-d0bd7c3037a2-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-rhzbg\" (UID: \"fecaea20-4e1a-4977-92e0-d0bd7c3037a2\") " pod="openshift-marketplace/marketplace-operator-79b997595-rhzbg" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.749915 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/fecaea20-4e1a-4977-92e0-d0bd7c3037a2-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-rhzbg\" (UID: \"fecaea20-4e1a-4977-92e0-d0bd7c3037a2\") " pod="openshift-marketplace/marketplace-operator-79b997595-rhzbg" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.749948 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6743e3ae-38d5-4ac1-8eea-7f0fbbe282da-apiservice-cert\") pod \"packageserver-d55dfcdfc-g8b88\" (UID: \"6743e3ae-38d5-4ac1-8eea-7f0fbbe282da\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g8b88" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.749974 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/cddf4720-7bab-43e0-989e-6cff2942a5fe-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-qd7d2\" (UID: \"cddf4720-7bab-43e0-989e-6cff2942a5fe\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qd7d2" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.750008 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6743e3ae-38d5-4ac1-8eea-7f0fbbe282da-webhook-cert\") pod \"packageserver-d55dfcdfc-g8b88\" (UID: \"6743e3ae-38d5-4ac1-8eea-7f0fbbe282da\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g8b88" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.750106 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/f42f4e15-4186-4a35-aed6-122f4518d222-signing-key\") pod \"service-ca-9c57cc56f-5btb4\" (UID: \"f42f4e15-4186-4a35-aed6-122f4518d222\") " pod="openshift-service-ca/service-ca-9c57cc56f-5btb4" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.750139 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/96a1e39b-0018-41b8-8d7c-9fa7697858e9-config-volume\") pod \"collect-profiles-29406510-f2lz7\" (UID: \"96a1e39b-0018-41b8-8d7c-9fa7697858e9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406510-f2lz7" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.750181 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e69469d9-e3b1-4afd-84de-9b7dafa8e9e4-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-2t46v\" (UID: \"e69469d9-e3b1-4afd-84de-9b7dafa8e9e4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-2t46v" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.750651 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/f42f4e15-4186-4a35-aed6-122f4518d222-signing-cabundle\") pod \"service-ca-9c57cc56f-5btb4\" (UID: \"f42f4e15-4186-4a35-aed6-122f4518d222\") " pod="openshift-service-ca/service-ca-9c57cc56f-5btb4" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.751932 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/96a1e39b-0018-41b8-8d7c-9fa7697858e9-config-volume\") pod \"collect-profiles-29406510-f2lz7\" (UID: \"96a1e39b-0018-41b8-8d7c-9fa7697858e9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406510-f2lz7" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.753273 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fecaea20-4e1a-4977-92e0-d0bd7c3037a2-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-rhzbg\" (UID: \"fecaea20-4e1a-4977-92e0-d0bd7c3037a2\") " pod="openshift-marketplace/marketplace-operator-79b997595-rhzbg" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.753713 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/f42f4e15-4186-4a35-aed6-122f4518d222-signing-key\") pod \"service-ca-9c57cc56f-5btb4\" (UID: \"f42f4e15-4186-4a35-aed6-122f4518d222\") " pod="openshift-service-ca/service-ca-9c57cc56f-5btb4" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.754312 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6743e3ae-38d5-4ac1-8eea-7f0fbbe282da-webhook-cert\") pod \"packageserver-d55dfcdfc-g8b88\" (UID: \"6743e3ae-38d5-4ac1-8eea-7f0fbbe282da\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g8b88" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.756183 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e69469d9-e3b1-4afd-84de-9b7dafa8e9e4-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-2t46v\" (UID: \"e69469d9-e3b1-4afd-84de-9b7dafa8e9e4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-2t46v" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.756234 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6743e3ae-38d5-4ac1-8eea-7f0fbbe282da-apiservice-cert\") pod \"packageserver-d55dfcdfc-g8b88\" (UID: \"6743e3ae-38d5-4ac1-8eea-7f0fbbe282da\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g8b88" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.757280 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/cddf4720-7bab-43e0-989e-6cff2942a5fe-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-qd7d2\" (UID: \"cddf4720-7bab-43e0-989e-6cff2942a5fe\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qd7d2" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.761168 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/fecaea20-4e1a-4977-92e0-d0bd7c3037a2-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-rhzbg\" (UID: \"fecaea20-4e1a-4977-92e0-d0bd7c3037a2\") " pod="openshift-marketplace/marketplace-operator-79b997595-rhzbg" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.765083 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2kp2\" (UniqueName: \"kubernetes.io/projected/bfdd16e4-1cb8-4178-af49-4db764abc507-kube-api-access-k2kp2\") pod \"control-plane-machine-set-operator-78cbb6b69f-85llw\" (UID: \"bfdd16e4-1cb8-4178-af49-4db764abc507\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-85llw" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.785965 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tbbpk\" (UniqueName: \"kubernetes.io/projected/7e204bf3-844c-460e-9f4a-7d78025a3b09-kube-api-access-tbbpk\") pod \"openshift-apiserver-operator-796bbdcf4f-kp5gm\" (UID: \"7e204bf3-844c-460e-9f4a-7d78025a3b09\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kp5gm" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.806894 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlnqf\" (UniqueName: \"kubernetes.io/projected/191743db-39d2-4f93-add6-471de6ccdede-kube-api-access-rlnqf\") pod \"apiserver-7bbb656c7d-zv6wk\" (UID: \"191743db-39d2-4f93-add6-471de6ccdede\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zv6wk" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.825703 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xb69\" (UniqueName: \"kubernetes.io/projected/ee4ae924-84bb-4b5c-9b6e-f03d6cba2ad3-kube-api-access-8xb69\") pod \"dns-operator-744455d44c-bpxtd\" (UID: \"ee4ae924-84bb-4b5c-9b6e-f03d6cba2ad3\") " pod="openshift-dns-operator/dns-operator-744455d44c-bpxtd" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.841932 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vk2xx" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.849386 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfk8z\" (UniqueName: \"kubernetes.io/projected/34ecc5fd-acf1-4748-af5d-d9af38bff6d3-kube-api-access-sfk8z\") pod \"apiserver-76f77b778f-rcsd7\" (UID: \"34ecc5fd-acf1-4748-af5d-d9af38bff6d3\") " pod="openshift-apiserver/apiserver-76f77b778f-rcsd7" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.869468 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e4bb8a0e-4090-4de8-b7b9-5b6fdc899cd0-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-4bfz8\" (UID: \"e4bb8a0e-4090-4de8-b7b9-5b6fdc899cd0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4bfz8" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.879258 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-bpxtd" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.887028 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xstcq\" (UniqueName: \"kubernetes.io/projected/b22a0046-b9c4-4b6e-ba05-1c453c998ff8-kube-api-access-xstcq\") pod \"openshift-config-operator-7777fb866f-tqr7m\" (UID: \"b22a0046-b9c4-4b6e-ba05-1c453c998ff8\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-tqr7m" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.911315 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lczb2\" (UniqueName: \"kubernetes.io/projected/fdba8af3-a255-470f-bffd-ced0a45e988d-kube-api-access-lczb2\") pod \"downloads-7954f5f757-bbm46\" (UID: \"fdba8af3-a255-470f-bffd-ced0a45e988d\") " pod="openshift-console/downloads-7954f5f757-bbm46" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.920007 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-6gm29"] Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.925509 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/19582088-032e-4b42-bc19-5080f7c97e13-bound-sa-token\") pod \"ingress-operator-5b745b69d9-9j7n6\" (UID: \"19582088-032e-4b42-bc19-5080f7c97e13\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9j7n6" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.949466 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gc56\" (UniqueName: \"kubernetes.io/projected/42edbb7d-0d06-4a5f-b481-7345a1b67227-kube-api-access-6gc56\") pod \"openshift-controller-manager-operator-756b6f6bc6-brs97\" (UID: \"42edbb7d-0d06-4a5f-b481-7345a1b67227\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-brs97" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.966428 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-brs97" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.973234 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vv4zx\" (UniqueName: \"kubernetes.io/projected/9f2aa3db-bf5b-47b3-bcec-5132916f0f7d-kube-api-access-vv4zx\") pod \"etcd-operator-b45778765-hkpdn\" (UID: \"9f2aa3db-bf5b-47b3-bcec-5132916f0f7d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hkpdn" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.986089 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-hkpdn" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.990251 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rj74b\" (UniqueName: \"kubernetes.io/projected/b6794cf7-9161-499d-ba18-74b95caaf3ae-kube-api-access-rj74b\") pod \"console-f9d7485db-2qb64\" (UID: \"b6794cf7-9161-499d-ba18-74b95caaf3ae\") " pod="openshift-console/console-f9d7485db-2qb64" Nov 29 04:40:13 crc kubenswrapper[4799]: I1129 04:40:13.996391 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4bfz8" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.006061 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvdj4\" (UniqueName: \"kubernetes.io/projected/a98eb357-7807-407c-a1e0-9d9b2b94e1fa-kube-api-access-kvdj4\") pod \"route-controller-manager-6576b87f9c-z7df5\" (UID: \"a98eb357-7807-407c-a1e0-9d9b2b94e1fa\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-z7df5" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.014502 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-bbm46" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.022170 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-85llw" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.027944 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrk57\" (UniqueName: \"kubernetes.io/projected/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-kube-api-access-jrk57\") pod \"oauth-openshift-558db77b4-vdxqp\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.028310 4799 request.go:700] Waited for 1.876570397s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-cluster-machine-approver/serviceaccounts/machine-approver-sa/token Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.034117 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kp5gm" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.048439 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vk2xx"] Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.055696 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqwlf\" (UniqueName: \"kubernetes.io/projected/9f245325-f6b0-4ffa-b423-63a47df33136-kube-api-access-mqwlf\") pod \"machine-approver-56656f9798-6jsdb\" (UID: \"9f245325-f6b0-4ffa-b423-63a47df33136\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6jsdb" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.059654 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-rcsd7" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.081442 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zv6wk" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.089281 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6t28\" (UniqueName: \"kubernetes.io/projected/938dcbc0-ed30-4200-9f6f-cbc1364cce88-kube-api-access-v6t28\") pod \"console-operator-58897d9998-svdzq\" (UID: \"938dcbc0-ed30-4200-9f6f-cbc1364cce88\") " pod="openshift-console-operator/console-operator-58897d9998-svdzq" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.091715 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.099399 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cdqg\" (UniqueName: \"kubernetes.io/projected/19582088-032e-4b42-bc19-5080f7c97e13-kube-api-access-4cdqg\") pod \"ingress-operator-5b745b69d9-9j7n6\" (UID: \"19582088-032e-4b42-bc19-5080f7c97e13\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9j7n6" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.102008 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-z7df5" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.104325 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-bpxtd"] Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.116007 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.129104 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6jsdb" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.131444 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.136596 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-tqr7m" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.153129 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 29 04:40:14 crc kubenswrapper[4799]: W1129 04:40:14.160094 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podee4ae924_84bb_4b5c_9b6e_f03d6cba2ad3.slice/crio-d3ae0928b140e407d17e199340989ba1cdeb68235dd89b46d8ac340312ccc415 WatchSource:0}: Error finding container d3ae0928b140e407d17e199340989ba1cdeb68235dd89b46d8ac340312ccc415: Status 404 returned error can't find the container with id d3ae0928b140e407d17e199340989ba1cdeb68235dd89b46d8ac340312ccc415 Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.207826 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8j7sv\" (UniqueName: \"kubernetes.io/projected/1c7126e3-85fe-4158-b92c-df2eb2717217-kube-api-access-8j7sv\") pod \"controller-manager-879f6c89f-ddl6h\" (UID: \"1c7126e3-85fe-4158-b92c-df2eb2717217\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ddl6h" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.221605 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/74f1d4e8-2506-43c8-bcbb-cb8bf28cfb35-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-xj4jw\" (UID: \"74f1d4e8-2506-43c8-bcbb-cb8bf28cfb35\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xj4jw" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.223502 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-brs97"] Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.232882 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7pq8\" (UniqueName: \"kubernetes.io/projected/36a3c13c-2670-4b33-a9fe-9071b35365fe-kube-api-access-r7pq8\") pod \"router-default-5444994796-jrhwp\" (UID: \"36a3c13c-2670-4b33-a9fe-9071b35365fe\") " pod="openshift-ingress/router-default-5444994796-jrhwp" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.250095 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mctqc\" (UniqueName: \"kubernetes.io/projected/50764326-a80d-4b7a-bfd3-2242f4d0e228-kube-api-access-mctqc\") pod \"authentication-operator-69f744f599-fnhgz\" (UID: \"50764326-a80d-4b7a-bfd3-2242f4d0e228\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fnhgz" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.250403 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.258265 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4bfz8"] Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.271168 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.277022 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-2qb64" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.277367 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-ddl6h" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.290700 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.299119 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4bfz8" event={"ID":"e4bb8a0e-4090-4de8-b7b9-5b6fdc899cd0","Type":"ContainerStarted","Data":"46299396bf7b3a68179bc88db45bb681d7482b0acb4758b82fd4f261d11fddeb"} Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.308200 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.311586 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-bpxtd" event={"ID":"ee4ae924-84bb-4b5c-9b6e-f03d6cba2ad3","Type":"ContainerStarted","Data":"d3ae0928b140e407d17e199340989ba1cdeb68235dd89b46d8ac340312ccc415"} Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.313738 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.324894 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vk2xx" event={"ID":"e86c829a-524d-4194-9ab2-717b21d7cd60","Type":"ContainerStarted","Data":"a94e76a68c4f30ee54580a1cb7fbb2ab28e8dd8d68801cc618a1826820cb22a8"} Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.328470 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6jsdb" event={"ID":"9f245325-f6b0-4ffa-b423-63a47df33136","Type":"ContainerStarted","Data":"f1c3bd22b01611c9585a8d1f14c356ef35f3f13c0a30fd24bb1c4b448606707b"} Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.330662 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.338638 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xj4jw" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.339617 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-jrhwp" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.352531 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.354323 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-wznnv" event={"ID":"daa3f105-2960-45a0-ab05-cb2cedea7e9f","Type":"ContainerStarted","Data":"c1309aaf4d27cf2192d4d8f05c479ea6ec5e0d351a0586e8afa90fc60cce4d4b"} Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.354381 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-wznnv" event={"ID":"daa3f105-2960-45a0-ab05-cb2cedea7e9f","Type":"ContainerStarted","Data":"638865ef5ccb60a407901cb192049c860159e074171d8296b31e536dce65dcac"} Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.355409 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9j7n6" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.356855 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-svdzq" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.357151 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-brs97" event={"ID":"42edbb7d-0d06-4a5f-b481-7345a1b67227","Type":"ContainerStarted","Data":"a428aef11c05744f283069cb292e90e14305a5250b54c91184300a85cab192ff"} Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.365495 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6gm29" event={"ID":"e73c6003-61f1-49b4-9275-14a55a35e186","Type":"ContainerStarted","Data":"1d61f99715862298faf51dd528df8a2c8f31c082db9dc7b35aa80cd8769742d2"} Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.365555 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6gm29" event={"ID":"e73c6003-61f1-49b4-9275-14a55a35e186","Type":"ContainerStarted","Data":"9807119ca4d839737543033f0dde3d37e26ad004b6d9352f09232c6359536c69"} Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.372521 4799 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.391808 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.412342 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.461630 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sffxt\" (UniqueName: \"kubernetes.io/projected/f42f4e15-4186-4a35-aed6-122f4518d222-kube-api-access-sffxt\") pod \"service-ca-9c57cc56f-5btb4\" (UID: \"f42f4e15-4186-4a35-aed6-122f4518d222\") " pod="openshift-service-ca/service-ca-9c57cc56f-5btb4" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.474371 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-5btb4" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.477057 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzcrd\" (UniqueName: \"kubernetes.io/projected/e69469d9-e3b1-4afd-84de-9b7dafa8e9e4-kube-api-access-lzcrd\") pod \"multus-admission-controller-857f4d67dd-2t46v\" (UID: \"e69469d9-e3b1-4afd-84de-9b7dafa8e9e4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-2t46v" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.527953 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8gdst\" (UniqueName: \"kubernetes.io/projected/f2baa9d1-df52-44c6-b62f-f39c16b70e34-kube-api-access-8gdst\") pod \"kube-storage-version-migrator-operator-b67b599dd-jq22x\" (UID: \"f2baa9d1-df52-44c6-b62f-f39c16b70e34\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jq22x" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.528389 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-fnhgz" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.536585 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7psw\" (UniqueName: \"kubernetes.io/projected/3a945cd2-2261-435a-8d3f-bacd5b808d26-kube-api-access-l7psw\") pod \"olm-operator-6b444d44fb-9krqq\" (UID: \"3a945cd2-2261-435a-8d3f-bacd5b808d26\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9krqq" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.550950 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nstwg\" (UniqueName: \"kubernetes.io/projected/6743e3ae-38d5-4ac1-8eea-7f0fbbe282da-kube-api-access-nstwg\") pod \"packageserver-d55dfcdfc-g8b88\" (UID: \"6743e3ae-38d5-4ac1-8eea-7f0fbbe282da\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g8b88" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.555063 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svwvw\" (UniqueName: \"kubernetes.io/projected/96a1e39b-0018-41b8-8d7c-9fa7697858e9-kube-api-access-svwvw\") pod \"collect-profiles-29406510-f2lz7\" (UID: \"96a1e39b-0018-41b8-8d7c-9fa7697858e9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406510-f2lz7" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.570014 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzpbl\" (UniqueName: \"kubernetes.io/projected/fecaea20-4e1a-4977-92e0-d0bd7c3037a2-kube-api-access-fzpbl\") pod \"marketplace-operator-79b997595-rhzbg\" (UID: \"fecaea20-4e1a-4977-92e0-d0bd7c3037a2\") " pod="openshift-marketplace/marketplace-operator-79b997595-rhzbg" Nov 29 04:40:14 crc kubenswrapper[4799]: W1129 04:40:14.580185 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod36a3c13c_2670_4b33_a9fe_9071b35365fe.slice/crio-1a6ef5548871461343337311faf144045050978ed226e6738cc2a7ec08cf203d WatchSource:0}: Error finding container 1a6ef5548871461343337311faf144045050978ed226e6738cc2a7ec08cf203d: Status 404 returned error can't find the container with id 1a6ef5548871461343337311faf144045050978ed226e6738cc2a7ec08cf203d Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.604079 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-bbm46"] Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.612451 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/aab927aa-d4cb-4b96-a5a6-bb14df8d4af8-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-hs96x\" (UID: \"aab927aa-d4cb-4b96-a5a6-bb14df8d4af8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hs96x" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.622981 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-hkpdn"] Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.630066 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvj24\" (UniqueName: \"kubernetes.io/projected/cddf4720-7bab-43e0-989e-6cff2942a5fe-kube-api-access-nvj24\") pod \"package-server-manager-789f6589d5-qd7d2\" (UID: \"cddf4720-7bab-43e0-989e-6cff2942a5fe\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qd7d2" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.634553 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98bj5\" (UniqueName: \"kubernetes.io/projected/2f1d8a27-12bd-4ac8-b23f-51f9f10fbcbf-kube-api-access-98bj5\") pod \"catalog-operator-68c6474976-qz5br\" (UID: \"2f1d8a27-12bd-4ac8-b23f-51f9f10fbcbf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qz5br" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.653983 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.662280 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jq22x" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.673555 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.673941 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qz5br" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.680402 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hs96x" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.689848 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.701202 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-z7df5"] Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.701259 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-85llw"] Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.713180 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.724136 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9krqq" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.728091 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-rcsd7"] Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.733315 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-2t46v" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.733813 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kp5gm"] Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.736072 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.748099 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g8b88" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.751184 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.780459 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406510-f2lz7" Nov 29 04:40:14 crc kubenswrapper[4799]: W1129 04:40:14.788146 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9f2aa3db_bf5b_47b3_bcec_5132916f0f7d.slice/crio-ab930f8dea365bfbc38fdc529004d5f153a0f6c4c03b9e3c58ae6968c2c55faf WatchSource:0}: Error finding container ab930f8dea365bfbc38fdc529004d5f153a0f6c4c03b9e3c58ae6968c2c55faf: Status 404 returned error can't find the container with id ab930f8dea365bfbc38fdc529004d5f153a0f6c4c03b9e3c58ae6968c2c55faf Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.789519 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-2qb64"] Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.789690 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-rhzbg" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.798470 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qd7d2" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.820647 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-zv6wk"] Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.836573 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-tqr7m"] Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.877051 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/48c7bbb5-36e2-4677-b259-d3da6302d415-registry-certificates\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.877120 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/48c7bbb5-36e2-4677-b259-d3da6302d415-ca-trust-extracted\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.877152 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/48c7bbb5-36e2-4677-b259-d3da6302d415-trusted-ca\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.877186 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrz2r\" (UniqueName: \"kubernetes.io/projected/482e1948-ee9d-401d-a590-ed64818891d3-kube-api-access-zrz2r\") pod \"migrator-59844c95c7-t4f4r\" (UID: \"482e1948-ee9d-401d-a590-ed64818891d3\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t4f4r" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.877216 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0f47a5c5-7c73-48ca-93a2-25d2973142d8-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-h4w54\" (UID: \"0f47a5c5-7c73-48ca-93a2-25d2973142d8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h4w54" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.877230 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdm9x\" (UniqueName: \"kubernetes.io/projected/0f47a5c5-7c73-48ca-93a2-25d2973142d8-kube-api-access-cdm9x\") pod \"cluster-image-registry-operator-dc59b4c8b-h4w54\" (UID: \"0f47a5c5-7c73-48ca-93a2-25d2973142d8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h4w54" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.877278 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvt7f\" (UniqueName: \"kubernetes.io/projected/48c7bbb5-36e2-4677-b259-d3da6302d415-kube-api-access-xvt7f\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.877306 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5h2lz\" (UniqueName: \"kubernetes.io/projected/c957ce17-f774-454a-a5e2-20f8ca7be2f0-kube-api-access-5h2lz\") pod \"machine-config-operator-74547568cd-nmvwb\" (UID: \"c957ce17-f774-454a-a5e2-20f8ca7be2f0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nmvwb" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.877337 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0f47a5c5-7c73-48ca-93a2-25d2973142d8-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-h4w54\" (UID: \"0f47a5c5-7c73-48ca-93a2-25d2973142d8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h4w54" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.877355 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c957ce17-f774-454a-a5e2-20f8ca7be2f0-images\") pod \"machine-config-operator-74547568cd-nmvwb\" (UID: \"c957ce17-f774-454a-a5e2-20f8ca7be2f0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nmvwb" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.877373 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/48c7bbb5-36e2-4677-b259-d3da6302d415-installation-pull-secrets\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.877390 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c957ce17-f774-454a-a5e2-20f8ca7be2f0-proxy-tls\") pod \"machine-config-operator-74547568cd-nmvwb\" (UID: \"c957ce17-f774-454a-a5e2-20f8ca7be2f0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nmvwb" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.877409 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cf3da23-866e-4182-bc35-021d2151f9f2-config\") pod \"service-ca-operator-777779d784-6rp4p\" (UID: \"6cf3da23-866e-4182-bc35-021d2151f9f2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6rp4p" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.877464 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c957ce17-f774-454a-a5e2-20f8ca7be2f0-auth-proxy-config\") pod \"machine-config-operator-74547568cd-nmvwb\" (UID: \"c957ce17-f774-454a-a5e2-20f8ca7be2f0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nmvwb" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.877496 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.877529 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/48c7bbb5-36e2-4677-b259-d3da6302d415-registry-tls\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.877554 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6cf3da23-866e-4182-bc35-021d2151f9f2-serving-cert\") pod \"service-ca-operator-777779d784-6rp4p\" (UID: \"6cf3da23-866e-4182-bc35-021d2151f9f2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6rp4p" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.877624 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/48c7bbb5-36e2-4677-b259-d3da6302d415-bound-sa-token\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.877652 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46bzd\" (UniqueName: \"kubernetes.io/projected/6cf3da23-866e-4182-bc35-021d2151f9f2-kube-api-access-46bzd\") pod \"service-ca-operator-777779d784-6rp4p\" (UID: \"6cf3da23-866e-4182-bc35-021d2151f9f2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6rp4p" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.877669 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/0f47a5c5-7c73-48ca-93a2-25d2973142d8-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-h4w54\" (UID: \"0f47a5c5-7c73-48ca-93a2-25d2973142d8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h4w54" Nov 29 04:40:14 crc kubenswrapper[4799]: E1129 04:40:14.879443 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 04:40:15.379427495 +0000 UTC m=+91.022357895 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2bbp8" (UID: "48c7bbb5-36e2-4677-b259-d3da6302d415") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.912353 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-ddl6h"] Nov 29 04:40:14 crc kubenswrapper[4799]: W1129 04:40:14.920125 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb22a0046_b9c4_4b6e_ba05_1c453c998ff8.slice/crio-31ea8893a7ae37ba5e060bd56a58c136446678055ac0112a5833a685c60d10be WatchSource:0}: Error finding container 31ea8893a7ae37ba5e060bd56a58c136446678055ac0112a5833a685c60d10be: Status 404 returned error can't find the container with id 31ea8893a7ae37ba5e060bd56a58c136446678055ac0112a5833a685c60d10be Nov 29 04:40:14 crc kubenswrapper[4799]: W1129 04:40:14.947277 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod34ecc5fd_acf1_4748_af5d_d9af38bff6d3.slice/crio-5ce863472597ad3c6c75c62b4bdb3c3dde50e395f44af3d232ad5fdd340d9dbf WatchSource:0}: Error finding container 5ce863472597ad3c6c75c62b4bdb3c3dde50e395f44af3d232ad5fdd340d9dbf: Status 404 returned error can't find the container with id 5ce863472597ad3c6c75c62b4bdb3c3dde50e395f44af3d232ad5fdd340d9dbf Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.978684 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xj4jw"] Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.979088 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 04:40:14 crc kubenswrapper[4799]: E1129 04:40:14.979359 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 04:40:15.47933827 +0000 UTC m=+91.122268670 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.979548 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/2aac0475-5e98-4db7-a0f4-6f128acb77c0-registration-dir\") pod \"csi-hostpathplugin-vh8lz\" (UID: \"2aac0475-5e98-4db7-a0f4-6f128acb77c0\") " pod="hostpath-provisioner/csi-hostpathplugin-vh8lz" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.979585 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0f47a5c5-7c73-48ca-93a2-25d2973142d8-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-h4w54\" (UID: \"0f47a5c5-7c73-48ca-93a2-25d2973142d8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h4w54" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.979635 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c957ce17-f774-454a-a5e2-20f8ca7be2f0-images\") pod \"machine-config-operator-74547568cd-nmvwb\" (UID: \"c957ce17-f774-454a-a5e2-20f8ca7be2f0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nmvwb" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.979657 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/48c7bbb5-36e2-4677-b259-d3da6302d415-installation-pull-secrets\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.979675 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c957ce17-f774-454a-a5e2-20f8ca7be2f0-proxy-tls\") pod \"machine-config-operator-74547568cd-nmvwb\" (UID: \"c957ce17-f774-454a-a5e2-20f8ca7be2f0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nmvwb" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.979692 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/2aac0475-5e98-4db7-a0f4-6f128acb77c0-mountpoint-dir\") pod \"csi-hostpathplugin-vh8lz\" (UID: \"2aac0475-5e98-4db7-a0f4-6f128acb77c0\") " pod="hostpath-provisioner/csi-hostpathplugin-vh8lz" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.979747 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cf3da23-866e-4182-bc35-021d2151f9f2-config\") pod \"service-ca-operator-777779d784-6rp4p\" (UID: \"6cf3da23-866e-4182-bc35-021d2151f9f2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6rp4p" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.979830 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c957ce17-f774-454a-a5e2-20f8ca7be2f0-auth-proxy-config\") pod \"machine-config-operator-74547568cd-nmvwb\" (UID: \"c957ce17-f774-454a-a5e2-20f8ca7be2f0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nmvwb" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.979889 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.979984 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/48c7bbb5-36e2-4677-b259-d3da6302d415-registry-tls\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.980129 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6cf3da23-866e-4182-bc35-021d2151f9f2-serving-cert\") pod \"service-ca-operator-777779d784-6rp4p\" (UID: \"6cf3da23-866e-4182-bc35-021d2151f9f2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6rp4p" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.981454 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cf3da23-866e-4182-bc35-021d2151f9f2-config\") pod \"service-ca-operator-777779d784-6rp4p\" (UID: \"6cf3da23-866e-4182-bc35-021d2151f9f2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6rp4p" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.981774 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c957ce17-f774-454a-a5e2-20f8ca7be2f0-auth-proxy-config\") pod \"machine-config-operator-74547568cd-nmvwb\" (UID: \"c957ce17-f774-454a-a5e2-20f8ca7be2f0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nmvwb" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.982146 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdt7m\" (UniqueName: \"kubernetes.io/projected/22463cfc-00a0-47f8-889a-35fea3fc4809-kube-api-access-bdt7m\") pod \"dns-default-vbffw\" (UID: \"22463cfc-00a0-47f8-889a-35fea3fc4809\") " pod="openshift-dns/dns-default-vbffw" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.982214 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/aef7387f-d081-4d53-bdb6-33f61918b318-node-bootstrap-token\") pod \"machine-config-server-2wtrf\" (UID: \"aef7387f-d081-4d53-bdb6-33f61918b318\") " pod="openshift-machine-config-operator/machine-config-server-2wtrf" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.982306 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhj5m\" (UniqueName: \"kubernetes.io/projected/aef7387f-d081-4d53-bdb6-33f61918b318-kube-api-access-zhj5m\") pod \"machine-config-server-2wtrf\" (UID: \"aef7387f-d081-4d53-bdb6-33f61918b318\") " pod="openshift-machine-config-operator/machine-config-server-2wtrf" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.982375 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/22463cfc-00a0-47f8-889a-35fea3fc4809-metrics-tls\") pod \"dns-default-vbffw\" (UID: \"22463cfc-00a0-47f8-889a-35fea3fc4809\") " pod="openshift-dns/dns-default-vbffw" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.982393 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2ghk\" (UniqueName: \"kubernetes.io/projected/f4a793d9-fe58-4613-bbd4-1550b95bd15a-kube-api-access-w2ghk\") pod \"ingress-canary-jjnx9\" (UID: \"f4a793d9-fe58-4613-bbd4-1550b95bd15a\") " pod="openshift-ingress-canary/ingress-canary-jjnx9" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.982579 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/48c7bbb5-36e2-4677-b259-d3da6302d415-bound-sa-token\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.982684 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46bzd\" (UniqueName: \"kubernetes.io/projected/6cf3da23-866e-4182-bc35-021d2151f9f2-kube-api-access-46bzd\") pod \"service-ca-operator-777779d784-6rp4p\" (UID: \"6cf3da23-866e-4182-bc35-021d2151f9f2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6rp4p" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.982780 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/0f47a5c5-7c73-48ca-93a2-25d2973142d8-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-h4w54\" (UID: \"0f47a5c5-7c73-48ca-93a2-25d2973142d8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h4w54" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.982847 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f4a793d9-fe58-4613-bbd4-1550b95bd15a-cert\") pod \"ingress-canary-jjnx9\" (UID: \"f4a793d9-fe58-4613-bbd4-1550b95bd15a\") " pod="openshift-ingress-canary/ingress-canary-jjnx9" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.982922 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/aef7387f-d081-4d53-bdb6-33f61918b318-certs\") pod \"machine-config-server-2wtrf\" (UID: \"aef7387f-d081-4d53-bdb6-33f61918b318\") " pod="openshift-machine-config-operator/machine-config-server-2wtrf" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.982972 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/48c7bbb5-36e2-4677-b259-d3da6302d415-ca-trust-extracted\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.982991 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/48c7bbb5-36e2-4677-b259-d3da6302d415-registry-certificates\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.983019 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/48c7bbb5-36e2-4677-b259-d3da6302d415-trusted-ca\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.983265 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrz2r\" (UniqueName: \"kubernetes.io/projected/482e1948-ee9d-401d-a590-ed64818891d3-kube-api-access-zrz2r\") pod \"migrator-59844c95c7-t4f4r\" (UID: \"482e1948-ee9d-401d-a590-ed64818891d3\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t4f4r" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.983308 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/2aac0475-5e98-4db7-a0f4-6f128acb77c0-plugins-dir\") pod \"csi-hostpathplugin-vh8lz\" (UID: \"2aac0475-5e98-4db7-a0f4-6f128acb77c0\") " pod="hostpath-provisioner/csi-hostpathplugin-vh8lz" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.983343 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0f47a5c5-7c73-48ca-93a2-25d2973142d8-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-h4w54\" (UID: \"0f47a5c5-7c73-48ca-93a2-25d2973142d8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h4w54" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.983384 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdm9x\" (UniqueName: \"kubernetes.io/projected/0f47a5c5-7c73-48ca-93a2-25d2973142d8-kube-api-access-cdm9x\") pod \"cluster-image-registry-operator-dc59b4c8b-h4w54\" (UID: \"0f47a5c5-7c73-48ca-93a2-25d2973142d8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h4w54" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.983400 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/2aac0475-5e98-4db7-a0f4-6f128acb77c0-socket-dir\") pod \"csi-hostpathplugin-vh8lz\" (UID: \"2aac0475-5e98-4db7-a0f4-6f128acb77c0\") " pod="hostpath-provisioner/csi-hostpathplugin-vh8lz" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.983462 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/22463cfc-00a0-47f8-889a-35fea3fc4809-config-volume\") pod \"dns-default-vbffw\" (UID: \"22463cfc-00a0-47f8-889a-35fea3fc4809\") " pod="openshift-dns/dns-default-vbffw" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.983653 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-js854\" (UniqueName: \"kubernetes.io/projected/2aac0475-5e98-4db7-a0f4-6f128acb77c0-kube-api-access-js854\") pod \"csi-hostpathplugin-vh8lz\" (UID: \"2aac0475-5e98-4db7-a0f4-6f128acb77c0\") " pod="hostpath-provisioner/csi-hostpathplugin-vh8lz" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.983728 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvt7f\" (UniqueName: \"kubernetes.io/projected/48c7bbb5-36e2-4677-b259-d3da6302d415-kube-api-access-xvt7f\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.991767 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/48c7bbb5-36e2-4677-b259-d3da6302d415-registry-tls\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.992554 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5h2lz\" (UniqueName: \"kubernetes.io/projected/c957ce17-f774-454a-a5e2-20f8ca7be2f0-kube-api-access-5h2lz\") pod \"machine-config-operator-74547568cd-nmvwb\" (UID: \"c957ce17-f774-454a-a5e2-20f8ca7be2f0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nmvwb" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.992632 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/2aac0475-5e98-4db7-a0f4-6f128acb77c0-csi-data-dir\") pod \"csi-hostpathplugin-vh8lz\" (UID: \"2aac0475-5e98-4db7-a0f4-6f128acb77c0\") " pod="hostpath-provisioner/csi-hostpathplugin-vh8lz" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.995355 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c957ce17-f774-454a-a5e2-20f8ca7be2f0-proxy-tls\") pod \"machine-config-operator-74547568cd-nmvwb\" (UID: \"c957ce17-f774-454a-a5e2-20f8ca7be2f0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nmvwb" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.996074 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6cf3da23-866e-4182-bc35-021d2151f9f2-serving-cert\") pod \"service-ca-operator-777779d784-6rp4p\" (UID: \"6cf3da23-866e-4182-bc35-021d2151f9f2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6rp4p" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.996175 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/48c7bbb5-36e2-4677-b259-d3da6302d415-installation-pull-secrets\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:14 crc kubenswrapper[4799]: I1129 04:40:14.999005 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-5btb4"] Nov 29 04:40:15 crc kubenswrapper[4799]: E1129 04:40:15.000449 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 04:40:15.500423924 +0000 UTC m=+91.143354334 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2bbp8" (UID: "48c7bbb5-36e2-4677-b259-d3da6302d415") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.000549 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/48c7bbb5-36e2-4677-b259-d3da6302d415-registry-certificates\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.000643 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/48c7bbb5-36e2-4677-b259-d3da6302d415-ca-trust-extracted\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.002985 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/48c7bbb5-36e2-4677-b259-d3da6302d415-trusted-ca\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.006054 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0f47a5c5-7c73-48ca-93a2-25d2973142d8-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-h4w54\" (UID: \"0f47a5c5-7c73-48ca-93a2-25d2973142d8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h4w54" Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.007533 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c957ce17-f774-454a-a5e2-20f8ca7be2f0-images\") pod \"machine-config-operator-74547568cd-nmvwb\" (UID: \"c957ce17-f774-454a-a5e2-20f8ca7be2f0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nmvwb" Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.007643 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/0f47a5c5-7c73-48ca-93a2-25d2973142d8-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-h4w54\" (UID: \"0f47a5c5-7c73-48ca-93a2-25d2973142d8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h4w54" Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.028030 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/48c7bbb5-36e2-4677-b259-d3da6302d415-bound-sa-token\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:15 crc kubenswrapper[4799]: W1129 04:40:15.034898 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod74f1d4e8_2506_43c8_bcbb_cb8bf28cfb35.slice/crio-0c5c052bf1817b1b465b112ef9f633ae926d6f27c2bde7020cccb844f6d3d261 WatchSource:0}: Error finding container 0c5c052bf1817b1b465b112ef9f633ae926d6f27c2bde7020cccb844f6d3d261: Status 404 returned error can't find the container with id 0c5c052bf1817b1b465b112ef9f633ae926d6f27c2bde7020cccb844f6d3d261 Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.043739 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdm9x\" (UniqueName: \"kubernetes.io/projected/0f47a5c5-7c73-48ca-93a2-25d2973142d8-kube-api-access-cdm9x\") pod \"cluster-image-registry-operator-dc59b4c8b-h4w54\" (UID: \"0f47a5c5-7c73-48ca-93a2-25d2973142d8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h4w54" Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.044209 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46bzd\" (UniqueName: \"kubernetes.io/projected/6cf3da23-866e-4182-bc35-021d2151f9f2-kube-api-access-46bzd\") pod \"service-ca-operator-777779d784-6rp4p\" (UID: \"6cf3da23-866e-4182-bc35-021d2151f9f2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6rp4p" Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.047997 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-6rp4p" Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.085668 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvt7f\" (UniqueName: \"kubernetes.io/projected/48c7bbb5-36e2-4677-b259-d3da6302d415-kube-api-access-xvt7f\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.096064 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.096244 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/2aac0475-5e98-4db7-a0f4-6f128acb77c0-plugins-dir\") pod \"csi-hostpathplugin-vh8lz\" (UID: \"2aac0475-5e98-4db7-a0f4-6f128acb77c0\") " pod="hostpath-provisioner/csi-hostpathplugin-vh8lz" Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.096272 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/2aac0475-5e98-4db7-a0f4-6f128acb77c0-socket-dir\") pod \"csi-hostpathplugin-vh8lz\" (UID: \"2aac0475-5e98-4db7-a0f4-6f128acb77c0\") " pod="hostpath-provisioner/csi-hostpathplugin-vh8lz" Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.096303 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/22463cfc-00a0-47f8-889a-35fea3fc4809-config-volume\") pod \"dns-default-vbffw\" (UID: \"22463cfc-00a0-47f8-889a-35fea3fc4809\") " pod="openshift-dns/dns-default-vbffw" Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.096344 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-js854\" (UniqueName: \"kubernetes.io/projected/2aac0475-5e98-4db7-a0f4-6f128acb77c0-kube-api-access-js854\") pod \"csi-hostpathplugin-vh8lz\" (UID: \"2aac0475-5e98-4db7-a0f4-6f128acb77c0\") " pod="hostpath-provisioner/csi-hostpathplugin-vh8lz" Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.096403 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/2aac0475-5e98-4db7-a0f4-6f128acb77c0-csi-data-dir\") pod \"csi-hostpathplugin-vh8lz\" (UID: \"2aac0475-5e98-4db7-a0f4-6f128acb77c0\") " pod="hostpath-provisioner/csi-hostpathplugin-vh8lz" Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.096426 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/2aac0475-5e98-4db7-a0f4-6f128acb77c0-registration-dir\") pod \"csi-hostpathplugin-vh8lz\" (UID: \"2aac0475-5e98-4db7-a0f4-6f128acb77c0\") " pod="hostpath-provisioner/csi-hostpathplugin-vh8lz" Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.096456 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/2aac0475-5e98-4db7-a0f4-6f128acb77c0-mountpoint-dir\") pod \"csi-hostpathplugin-vh8lz\" (UID: \"2aac0475-5e98-4db7-a0f4-6f128acb77c0\") " pod="hostpath-provisioner/csi-hostpathplugin-vh8lz" Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.096513 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdt7m\" (UniqueName: \"kubernetes.io/projected/22463cfc-00a0-47f8-889a-35fea3fc4809-kube-api-access-bdt7m\") pod \"dns-default-vbffw\" (UID: \"22463cfc-00a0-47f8-889a-35fea3fc4809\") " pod="openshift-dns/dns-default-vbffw" Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.096550 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/aef7387f-d081-4d53-bdb6-33f61918b318-node-bootstrap-token\") pod \"machine-config-server-2wtrf\" (UID: \"aef7387f-d081-4d53-bdb6-33f61918b318\") " pod="openshift-machine-config-operator/machine-config-server-2wtrf" Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.096577 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/22463cfc-00a0-47f8-889a-35fea3fc4809-metrics-tls\") pod \"dns-default-vbffw\" (UID: \"22463cfc-00a0-47f8-889a-35fea3fc4809\") " pod="openshift-dns/dns-default-vbffw" Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.096602 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2ghk\" (UniqueName: \"kubernetes.io/projected/f4a793d9-fe58-4613-bbd4-1550b95bd15a-kube-api-access-w2ghk\") pod \"ingress-canary-jjnx9\" (UID: \"f4a793d9-fe58-4613-bbd4-1550b95bd15a\") " pod="openshift-ingress-canary/ingress-canary-jjnx9" Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.096627 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhj5m\" (UniqueName: \"kubernetes.io/projected/aef7387f-d081-4d53-bdb6-33f61918b318-kube-api-access-zhj5m\") pod \"machine-config-server-2wtrf\" (UID: \"aef7387f-d081-4d53-bdb6-33f61918b318\") " pod="openshift-machine-config-operator/machine-config-server-2wtrf" Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.096655 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f4a793d9-fe58-4613-bbd4-1550b95bd15a-cert\") pod \"ingress-canary-jjnx9\" (UID: \"f4a793d9-fe58-4613-bbd4-1550b95bd15a\") " pod="openshift-ingress-canary/ingress-canary-jjnx9" Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.096675 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/aef7387f-d081-4d53-bdb6-33f61918b318-certs\") pod \"machine-config-server-2wtrf\" (UID: \"aef7387f-d081-4d53-bdb6-33f61918b318\") " pod="openshift-machine-config-operator/machine-config-server-2wtrf" Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.097445 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/2aac0475-5e98-4db7-a0f4-6f128acb77c0-registration-dir\") pod \"csi-hostpathplugin-vh8lz\" (UID: \"2aac0475-5e98-4db7-a0f4-6f128acb77c0\") " pod="hostpath-provisioner/csi-hostpathplugin-vh8lz" Nov 29 04:40:15 crc kubenswrapper[4799]: E1129 04:40:15.097542 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 04:40:15.597524897 +0000 UTC m=+91.240455297 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.097579 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/2aac0475-5e98-4db7-a0f4-6f128acb77c0-plugins-dir\") pod \"csi-hostpathplugin-vh8lz\" (UID: \"2aac0475-5e98-4db7-a0f4-6f128acb77c0\") " pod="hostpath-provisioner/csi-hostpathplugin-vh8lz" Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.097617 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/2aac0475-5e98-4db7-a0f4-6f128acb77c0-socket-dir\") pod \"csi-hostpathplugin-vh8lz\" (UID: \"2aac0475-5e98-4db7-a0f4-6f128acb77c0\") " pod="hostpath-provisioner/csi-hostpathplugin-vh8lz" Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.098174 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/22463cfc-00a0-47f8-889a-35fea3fc4809-config-volume\") pod \"dns-default-vbffw\" (UID: \"22463cfc-00a0-47f8-889a-35fea3fc4809\") " pod="openshift-dns/dns-default-vbffw" Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.098331 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/2aac0475-5e98-4db7-a0f4-6f128acb77c0-csi-data-dir\") pod \"csi-hostpathplugin-vh8lz\" (UID: \"2aac0475-5e98-4db7-a0f4-6f128acb77c0\") " pod="hostpath-provisioner/csi-hostpathplugin-vh8lz" Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.100290 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/aef7387f-d081-4d53-bdb6-33f61918b318-certs\") pod \"machine-config-server-2wtrf\" (UID: \"aef7387f-d081-4d53-bdb6-33f61918b318\") " pod="openshift-machine-config-operator/machine-config-server-2wtrf" Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.100977 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/2aac0475-5e98-4db7-a0f4-6f128acb77c0-mountpoint-dir\") pod \"csi-hostpathplugin-vh8lz\" (UID: \"2aac0475-5e98-4db7-a0f4-6f128acb77c0\") " pod="hostpath-provisioner/csi-hostpathplugin-vh8lz" Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.115035 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-fnhgz"] Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.118065 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/22463cfc-00a0-47f8-889a-35fea3fc4809-metrics-tls\") pod \"dns-default-vbffw\" (UID: \"22463cfc-00a0-47f8-889a-35fea3fc4809\") " pod="openshift-dns/dns-default-vbffw" Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.119299 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5h2lz\" (UniqueName: \"kubernetes.io/projected/c957ce17-f774-454a-a5e2-20f8ca7be2f0-kube-api-access-5h2lz\") pod \"machine-config-operator-74547568cd-nmvwb\" (UID: \"c957ce17-f774-454a-a5e2-20f8ca7be2f0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nmvwb" Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.119614 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/aef7387f-d081-4d53-bdb6-33f61918b318-node-bootstrap-token\") pod \"machine-config-server-2wtrf\" (UID: \"aef7387f-d081-4d53-bdb6-33f61918b318\") " pod="openshift-machine-config-operator/machine-config-server-2wtrf" Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.124652 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f4a793d9-fe58-4613-bbd4-1550b95bd15a-cert\") pod \"ingress-canary-jjnx9\" (UID: \"f4a793d9-fe58-4613-bbd4-1550b95bd15a\") " pod="openshift-ingress-canary/ingress-canary-jjnx9" Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.161278 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-vdxqp"] Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.197901 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:15 crc kubenswrapper[4799]: E1129 04:40:15.198288 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 04:40:15.698271564 +0000 UTC m=+91.341201964 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2bbp8" (UID: "48c7bbb5-36e2-4677-b259-d3da6302d415") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.200125 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0f47a5c5-7c73-48ca-93a2-25d2973142d8-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-h4w54\" (UID: \"0f47a5c5-7c73-48ca-93a2-25d2973142d8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h4w54" Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.201658 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrz2r\" (UniqueName: \"kubernetes.io/projected/482e1948-ee9d-401d-a590-ed64818891d3-kube-api-access-zrz2r\") pod \"migrator-59844c95c7-t4f4r\" (UID: \"482e1948-ee9d-401d-a590-ed64818891d3\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t4f4r" Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.214489 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-js854\" (UniqueName: \"kubernetes.io/projected/2aac0475-5e98-4db7-a0f4-6f128acb77c0-kube-api-access-js854\") pod \"csi-hostpathplugin-vh8lz\" (UID: \"2aac0475-5e98-4db7-a0f4-6f128acb77c0\") " pod="hostpath-provisioner/csi-hostpathplugin-vh8lz" Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.223723 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-svdzq"] Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.224931 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-9j7n6"] Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.236467 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2ghk\" (UniqueName: \"kubernetes.io/projected/f4a793d9-fe58-4613-bbd4-1550b95bd15a-kube-api-access-w2ghk\") pod \"ingress-canary-jjnx9\" (UID: \"f4a793d9-fe58-4613-bbd4-1550b95bd15a\") " pod="openshift-ingress-canary/ingress-canary-jjnx9" Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.243426 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhj5m\" (UniqueName: \"kubernetes.io/projected/aef7387f-d081-4d53-bdb6-33f61918b318-kube-api-access-zhj5m\") pod \"machine-config-server-2wtrf\" (UID: \"aef7387f-d081-4d53-bdb6-33f61918b318\") " pod="openshift-machine-config-operator/machine-config-server-2wtrf" Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.249916 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdt7m\" (UniqueName: \"kubernetes.io/projected/22463cfc-00a0-47f8-889a-35fea3fc4809-kube-api-access-bdt7m\") pod \"dns-default-vbffw\" (UID: \"22463cfc-00a0-47f8-889a-35fea3fc4809\") " pod="openshift-dns/dns-default-vbffw" Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.291657 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nmvwb" Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.299070 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 04:40:15 crc kubenswrapper[4799]: E1129 04:40:15.299566 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 04:40:15.799550305 +0000 UTC m=+91.442480705 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.300293 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jq22x"] Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.303340 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t4f4r" Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.359658 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h4w54" Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.380118 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hs96x"] Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.390569 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-bbm46" event={"ID":"fdba8af3-a255-470f-bffd-ced0a45e988d","Type":"ContainerStarted","Data":"be50bc2b781693858ce30faa85e78d8b1390ee1b0996e21bfc786318378fa07c"} Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.393489 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-5btb4" event={"ID":"f42f4e15-4186-4a35-aed6-122f4518d222","Type":"ContainerStarted","Data":"6d069970a8e0d553c9b6ba36e830d40b2387e8e4700354b37364ff94bab64c59"} Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.395971 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kp5gm" event={"ID":"7e204bf3-844c-460e-9f4a-7d78025a3b09","Type":"ContainerStarted","Data":"39da3e9e5a5a1a8f2805a38688a6aa39c3501c6e6f7d0b18c6af8cc5f0288178"} Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.398021 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zv6wk" event={"ID":"191743db-39d2-4f93-add6-471de6ccdede","Type":"ContainerStarted","Data":"bcbfbe7746296e8c9bdbb55b54b96d2556a3d94fb85475227ef5d107dc013b8d"} Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.399489 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-2qb64" event={"ID":"b6794cf7-9161-499d-ba18-74b95caaf3ae","Type":"ContainerStarted","Data":"714279d9206d684d361016c0a993451532ec8a99d097a25966400cbd7b1d7d7d"} Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.400316 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:15 crc kubenswrapper[4799]: E1129 04:40:15.400581 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 04:40:15.900569989 +0000 UTC m=+91.543500389 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2bbp8" (UID: "48c7bbb5-36e2-4677-b259-d3da6302d415") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.405914 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-fnhgz" event={"ID":"50764326-a80d-4b7a-bfd3-2242f4d0e228","Type":"ContainerStarted","Data":"a493c945b33126a6fc3de2a53b07076b8c3ee76bcb6d48faf68fc9eda71078b7"} Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.406057 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-jjnx9" Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.411518 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-ddl6h" event={"ID":"1c7126e3-85fe-4158-b92c-df2eb2717217","Type":"ContainerStarted","Data":"d81239a6dc2d3950eeaf981a7e488978b8cc60f1af5358190a59f24c87edcd98"} Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.412821 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-tqr7m" event={"ID":"b22a0046-b9c4-4b6e-ba05-1c453c998ff8","Type":"ContainerStarted","Data":"31ea8893a7ae37ba5e060bd56a58c136446678055ac0112a5833a685c60d10be"} Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.419865 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-brs97" event={"ID":"42edbb7d-0d06-4a5f-b481-7345a1b67227","Type":"ContainerStarted","Data":"3ba5dea1f6d18eea59b5f568bca7706166c73a4f9ba6630671d6a5bee98b1f01"} Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.423173 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-2wtrf" Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.423236 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-hkpdn" event={"ID":"9f2aa3db-bf5b-47b3-bcec-5132916f0f7d","Type":"ContainerStarted","Data":"ab930f8dea365bfbc38fdc529004d5f153a0f6c4c03b9e3c58ae6968c2c55faf"} Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.424621 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-bpxtd" event={"ID":"ee4ae924-84bb-4b5c-9b6e-f03d6cba2ad3","Type":"ContainerStarted","Data":"3719543fd752f441c2919d879222fb067bcee32f89320111f2cfa7b37b8defb1"} Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.427913 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-rcsd7" event={"ID":"34ecc5fd-acf1-4748-af5d-d9af38bff6d3","Type":"ContainerStarted","Data":"5ce863472597ad3c6c75c62b4bdb3c3dde50e395f44af3d232ad5fdd340d9dbf"} Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.433175 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-vbffw" Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.447596 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-vh8lz" Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.459117 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6gm29" event={"ID":"e73c6003-61f1-49b4-9275-14a55a35e186","Type":"ContainerStarted","Data":"4b742ef2656581db594694bc07cdb49e4c284ca744a2cc3ba4c61431f2532ba7"} Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.464624 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-jrhwp" event={"ID":"36a3c13c-2670-4b33-a9fe-9071b35365fe","Type":"ContainerStarted","Data":"1a6ef5548871461343337311faf144045050978ed226e6738cc2a7ec08cf203d"} Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.475637 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xj4jw" event={"ID":"74f1d4e8-2506-43c8-bcbb-cb8bf28cfb35","Type":"ContainerStarted","Data":"0c5c052bf1817b1b465b112ef9f633ae926d6f27c2bde7020cccb844f6d3d261"} Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.477047 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-z7df5" event={"ID":"a98eb357-7807-407c-a1e0-9d9b2b94e1fa","Type":"ContainerStarted","Data":"80ed6fbe5b745575492889847c82f43bf37fd9315ccbb3b69bcb86b8e12d542a"} Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.477807 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" event={"ID":"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d","Type":"ContainerStarted","Data":"1ad557578cb04dc2f7758cf89412842a705604f1a60338baf488e205cbcbb90a"} Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.479477 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vk2xx" event={"ID":"e86c829a-524d-4194-9ab2-717b21d7cd60","Type":"ContainerStarted","Data":"df92e2a57f2213c2ed13b44326034223b277c091db154c8a6e27b5bd98eeae17"} Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.479501 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vk2xx" event={"ID":"e86c829a-524d-4194-9ab2-717b21d7cd60","Type":"ContainerStarted","Data":"f81d814dc923ed9c870668d190f7e2e76828c5d3c5cb7ba205bb2684056549b0"} Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.483930 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-svdzq" event={"ID":"938dcbc0-ed30-4200-9f6f-cbc1364cce88","Type":"ContainerStarted","Data":"2ca883b0d3122acff6f396f06ceddeacda6e249090abb7c1e6a38bd97dd986d6"} Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.484759 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9j7n6" event={"ID":"19582088-032e-4b42-bc19-5080f7c97e13","Type":"ContainerStarted","Data":"21206dcf2c3dc4b394c6ba97af747bd9b3c223c488a6305a1b02dfeff94f7a2b"} Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.485603 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-85llw" event={"ID":"bfdd16e4-1cb8-4178-af49-4db764abc507","Type":"ContainerStarted","Data":"fe3ce044d81de2ad9ef06e89f2b413e9eebca774110a4c134109e0a8c5af01b4"} Nov 29 04:40:15 crc kubenswrapper[4799]: W1129 04:40:15.490480 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf2baa9d1_df52_44c6_b62f_f39c16b70e34.slice/crio-8e048e2dff012652d8592a7f08910b5723d72672f63be23b29f04048f0688ee7 WatchSource:0}: Error finding container 8e048e2dff012652d8592a7f08910b5723d72672f63be23b29f04048f0688ee7: Status 404 returned error can't find the container with id 8e048e2dff012652d8592a7f08910b5723d72672f63be23b29f04048f0688ee7 Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.505533 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 04:40:15 crc kubenswrapper[4799]: E1129 04:40:15.505924 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 04:40:16.005691819 +0000 UTC m=+91.648622219 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.506382 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:15 crc kubenswrapper[4799]: E1129 04:40:15.507250 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 04:40:16.007235009 +0000 UTC m=+91.650165409 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2bbp8" (UID: "48c7bbb5-36e2-4677-b259-d3da6302d415") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.607151 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 04:40:15 crc kubenswrapper[4799]: E1129 04:40:15.610143 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 04:40:16.110123341 +0000 UTC m=+91.753053741 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.705115 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qz5br"] Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.715936 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:15 crc kubenswrapper[4799]: E1129 04:40:15.718562 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 04:40:16.218541916 +0000 UTC m=+91.861472336 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2bbp8" (UID: "48c7bbb5-36e2-4677-b259-d3da6302d415") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.751313 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-2t46v"] Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.788209 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g8b88"] Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.817206 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 04:40:15 crc kubenswrapper[4799]: E1129 04:40:15.817565 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 04:40:16.317548359 +0000 UTC m=+91.960478749 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.868328 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-wznnv" podStartSLOduration=66.868305507 podStartE2EDuration="1m6.868305507s" podCreationTimestamp="2025-11-29 04:39:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:15.827762092 +0000 UTC m=+91.470692492" watchObservedRunningTime="2025-11-29 04:40:15.868305507 +0000 UTC m=+91.511235907" Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.870487 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406510-f2lz7"] Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.882641 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rhzbg"] Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.888167 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qd7d2"] Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.891101 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9krqq"] Nov 29 04:40:15 crc kubenswrapper[4799]: I1129 04:40:15.919977 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:15 crc kubenswrapper[4799]: E1129 04:40:15.920348 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 04:40:16.420336988 +0000 UTC m=+92.063267388 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2bbp8" (UID: "48c7bbb5-36e2-4677-b259-d3da6302d415") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.005594 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-t4f4r"] Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.037566 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 04:40:16 crc kubenswrapper[4799]: E1129 04:40:16.037844 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 04:40:16.537820917 +0000 UTC m=+92.180751317 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.042266 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-6rp4p"] Nov 29 04:40:16 crc kubenswrapper[4799]: W1129 04:40:16.090034 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod96a1e39b_0018_41b8_8d7c_9fa7697858e9.slice/crio-73a3ef287032122637ca8b076df9af0da790e3bd7b0ca554cf8f473d569aff12 WatchSource:0}: Error finding container 73a3ef287032122637ca8b076df9af0da790e3bd7b0ca554cf8f473d569aff12: Status 404 returned error can't find the container with id 73a3ef287032122637ca8b076df9af0da790e3bd7b0ca554cf8f473d569aff12 Nov 29 04:40:16 crc kubenswrapper[4799]: W1129 04:40:16.090444 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfecaea20_4e1a_4977_92e0_d0bd7c3037a2.slice/crio-ba7cb2066e7577d7203ab116da03c50c95efc045b17699d2cbf3194bf3fa70a8 WatchSource:0}: Error finding container ba7cb2066e7577d7203ab116da03c50c95efc045b17699d2cbf3194bf3fa70a8: Status 404 returned error can't find the container with id ba7cb2066e7577d7203ab116da03c50c95efc045b17699d2cbf3194bf3fa70a8 Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.139052 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:16 crc kubenswrapper[4799]: E1129 04:40:16.139834 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 04:40:16.639821307 +0000 UTC m=+92.282751707 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2bbp8" (UID: "48c7bbb5-36e2-4677-b259-d3da6302d415") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.245208 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 04:40:16 crc kubenswrapper[4799]: E1129 04:40:16.245718 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 04:40:16.745704006 +0000 UTC m=+92.388634406 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.344662 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-jrhwp" Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.344714 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h4w54"] Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.347287 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:16 crc kubenswrapper[4799]: E1129 04:40:16.347566 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 04:40:16.847554192 +0000 UTC m=+92.490484592 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2bbp8" (UID: "48c7bbb5-36e2-4677-b259-d3da6302d415") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.356606 4799 patch_prober.go:28] interesting pod/router-default-5444994796-jrhwp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 29 04:40:16 crc kubenswrapper[4799]: [-]has-synced failed: reason withheld Nov 29 04:40:16 crc kubenswrapper[4799]: [+]process-running ok Nov 29 04:40:16 crc kubenswrapper[4799]: healthz check failed Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.356653 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jrhwp" podUID="36a3c13c-2670-4b33-a9fe-9071b35365fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.448348 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 04:40:16 crc kubenswrapper[4799]: E1129 04:40:16.448692 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 04:40:16.948676248 +0000 UTC m=+92.591606648 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.529209 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-6rp4p" event={"ID":"6cf3da23-866e-4182-bc35-021d2151f9f2","Type":"ContainerStarted","Data":"1035ae8b8fc5e595b8837c1ec276cf0a43b93de1bbfbd3ebd80e70505bf67fb8"} Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.549937 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:16 crc kubenswrapper[4799]: E1129 04:40:16.550505 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 04:40:17.050493013 +0000 UTC m=+92.693423413 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2bbp8" (UID: "48c7bbb5-36e2-4677-b259-d3da6302d415") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.591368 4799 generic.go:334] "Generic (PLEG): container finished" podID="b22a0046-b9c4-4b6e-ba05-1c453c998ff8" containerID="bfac174541c22f35ded85aeb42368913f78f0b3ddc7accf6ba28a3a9a68732b0" exitCode=0 Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.591755 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-tqr7m" event={"ID":"b22a0046-b9c4-4b6e-ba05-1c453c998ff8","Type":"ContainerDied","Data":"bfac174541c22f35ded85aeb42368913f78f0b3ddc7accf6ba28a3a9a68732b0"} Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.608227 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kp5gm" event={"ID":"7e204bf3-844c-460e-9f4a-7d78025a3b09","Type":"ContainerStarted","Data":"957d4395e127967bb5948d5edb27cb49a5fc880579b132a438a92a780013d9d5"} Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.644174 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29406510-f2lz7" event={"ID":"96a1e39b-0018-41b8-8d7c-9fa7697858e9","Type":"ContainerStarted","Data":"73a3ef287032122637ca8b076df9af0da790e3bd7b0ca554cf8f473d569aff12"} Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.656709 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 04:40:16 crc kubenswrapper[4799]: E1129 04:40:16.656848 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 04:40:17.156821274 +0000 UTC m=+92.799751674 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.656991 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:16 crc kubenswrapper[4799]: E1129 04:40:16.657698 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 04:40:17.157680377 +0000 UTC m=+92.800610777 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2bbp8" (UID: "48c7bbb5-36e2-4677-b259-d3da6302d415") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.717961 4799 patch_prober.go:28] interesting pod/downloads-7954f5f757-bbm46 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.28:8080/\": dial tcp 10.217.0.28:8080: connect: connection refused" start-of-body= Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.718396 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-bbm46" podUID="fdba8af3-a255-470f-bffd-ced0a45e988d" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.28:8080/\": dial tcp 10.217.0.28:8080: connect: connection refused" Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.762480 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 04:40:16 crc kubenswrapper[4799]: E1129 04:40:16.764079 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 04:40:17.264061449 +0000 UTC m=+92.906991839 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.828972 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6jsdb" event={"ID":"9f245325-f6b0-4ffa-b423-63a47df33136","Type":"ContainerStarted","Data":"86f2c1c9ae1455b34f5250414a1bbbe383364602d40d986b877f2a2478e84e93"} Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.829016 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-ddl6h" Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.829028 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g8b88" event={"ID":"6743e3ae-38d5-4ac1-8eea-7f0fbbe282da","Type":"ContainerStarted","Data":"d3fd8e6316e5c86cc54e49d4aedbcd9202489e8842e45060ad6c74fb49a2f69e"} Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.829040 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t4f4r" event={"ID":"482e1948-ee9d-401d-a590-ed64818891d3","Type":"ContainerStarted","Data":"07194f011f0d92844ca2938e1adf81ae52f48bc6dddd6db0c4ec2eeff983d068"} Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.829055 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-bbm46" Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.829075 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-vh8lz"] Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.829108 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-ddl6h" Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.829118 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-z7df5" Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.829131 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-ddl6h" event={"ID":"1c7126e3-85fe-4158-b92c-df2eb2717217","Type":"ContainerStarted","Data":"a54d5d6a40e2a60ac476cb198cfdccb043ded7feb9925bd82143833a3786baed"} Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.829140 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-2qb64" event={"ID":"b6794cf7-9161-499d-ba18-74b95caaf3ae","Type":"ContainerStarted","Data":"de4da2900ad12d1a9926d491dac701179128832d39109e72fcb7ee0cad10ff84"} Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.829151 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-85llw" event={"ID":"bfdd16e4-1cb8-4178-af49-4db764abc507","Type":"ContainerStarted","Data":"beb9582ff3999189b7b5325c113d825461127d501fff46691ab4a59d74a22c00"} Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.829162 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-bbm46" event={"ID":"fdba8af3-a255-470f-bffd-ced0a45e988d","Type":"ContainerStarted","Data":"f0b835372ec72be9ae08c1f36085a5ed65f54b81004d2a98cb57ecd90ad60ddd"} Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.829173 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.829189 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-z7df5" event={"ID":"a98eb357-7807-407c-a1e0-9d9b2b94e1fa","Type":"ContainerStarted","Data":"eaaf73a2d39291bdd9d3c42cc35e90d2003d3cce6a94fc4dd29c313a9ae7b133"} Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.829203 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-2wtrf" event={"ID":"aef7387f-d081-4d53-bdb6-33f61918b318","Type":"ContainerStarted","Data":"f9e389ea7368d2911cb34983ade1e46d2f103ce50636769ee9f4e6c8c3a2a457"} Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.829213 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4bfz8" event={"ID":"e4bb8a0e-4090-4de8-b7b9-5b6fdc899cd0","Type":"ContainerStarted","Data":"12c1248c7b131aa078aee26c522f12350e77afc9eca0fef8e7f2036b5b4bd031"} Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.829224 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hs96x" event={"ID":"aab927aa-d4cb-4b96-a5a6-bb14df8d4af8","Type":"ContainerStarted","Data":"e8f90b5791eb468e71c55c2b506d73bd1aa6a97d6a231133689b78f2418c7ca2"} Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.829236 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jq22x" event={"ID":"f2baa9d1-df52-44c6-b62f-f39c16b70e34","Type":"ContainerStarted","Data":"8e048e2dff012652d8592a7f08910b5723d72672f63be23b29f04048f0688ee7"} Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.837904 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9krqq" event={"ID":"3a945cd2-2261-435a-8d3f-bacd5b808d26","Type":"ContainerStarted","Data":"d16c9443debe251f07cbfa75aa67a3a81e08ea7505ae6cb23817d301179a059d"} Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.842033 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-rhzbg" event={"ID":"fecaea20-4e1a-4977-92e0-d0bd7c3037a2","Type":"ContainerStarted","Data":"ba7cb2066e7577d7203ab116da03c50c95efc045b17699d2cbf3194bf3fa70a8"} Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.850280 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qd7d2" event={"ID":"cddf4720-7bab-43e0-989e-6cff2942a5fe","Type":"ContainerStarted","Data":"55d8bd34443468d8678706e868efbbbd4f884c198b44f567b83a5a4cf139e934"} Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.853729 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-jrhwp" event={"ID":"36a3c13c-2670-4b33-a9fe-9071b35365fe","Type":"ContainerStarted","Data":"ed2ceb877cf2c24bb6e6a61fe5888874a6d3c3515de025ba265602a585d47050"} Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.854485 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-bbm46" podStartSLOduration=67.85446968 podStartE2EDuration="1m7.85446968s" podCreationTimestamp="2025-11-29 04:39:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:16.801370801 +0000 UTC m=+92.444301191" watchObservedRunningTime="2025-11-29 04:40:16.85446968 +0000 UTC m=+92.497400080" Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.858642 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-jrhwp" podStartSLOduration=67.858632397 podStartE2EDuration="1m7.858632397s" podCreationTimestamp="2025-11-29 04:39:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:16.85367955 +0000 UTC m=+92.496609950" watchObservedRunningTime="2025-11-29 04:40:16.858632397 +0000 UTC m=+92.501562797" Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.859978 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-5btb4" event={"ID":"f42f4e15-4186-4a35-aed6-122f4518d222","Type":"ContainerStarted","Data":"75c347bbfd790db2320988a5b2addbb5a2506dfbab29d2967e83c636fbf93ce7"} Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.863043 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-2t46v" event={"ID":"e69469d9-e3b1-4afd-84de-9b7dafa8e9e4","Type":"ContainerStarted","Data":"05fea2e66513140db7928822747c37e3d9ec66b72e255d0c3424147509a08261"} Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.866985 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:16 crc kubenswrapper[4799]: E1129 04:40:16.869531 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 04:40:17.369514728 +0000 UTC m=+93.012445128 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2bbp8" (UID: "48c7bbb5-36e2-4677-b259-d3da6302d415") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.889704 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qz5br" event={"ID":"2f1d8a27-12bd-4ac8-b23f-51f9f10fbcbf","Type":"ContainerStarted","Data":"3a5bbebc804a6ebc2e63c47acbee99a4750000c2340fbf161adf98eccc0ba02f"} Nov 29 04:40:16 crc kubenswrapper[4799]: W1129 04:40:16.923858 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2aac0475_5e98_4db7_a0f4_6f128acb77c0.slice/crio-d125f773b479bbf24ca318124b7074e712128e7a9db78dcc6e36df17797e0930 WatchSource:0}: Error finding container d125f773b479bbf24ca318124b7074e712128e7a9db78dcc6e36df17797e0930: Status 404 returned error can't find the container with id d125f773b479bbf24ca318124b7074e712128e7a9db78dcc6e36df17797e0930 Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.935911 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vk2xx" podStartSLOduration=67.935897619 podStartE2EDuration="1m7.935897619s" podCreationTimestamp="2025-11-29 04:39:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:16.935120289 +0000 UTC m=+92.578050689" watchObservedRunningTime="2025-11-29 04:40:16.935897619 +0000 UTC m=+92.578828019" Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.936523 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-brs97" podStartSLOduration=67.936518475 podStartE2EDuration="1m7.936518475s" podCreationTimestamp="2025-11-29 04:39:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:16.889033231 +0000 UTC m=+92.531963631" watchObservedRunningTime="2025-11-29 04:40:16.936518475 +0000 UTC m=+92.579448875" Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.970772 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-z7df5" Nov 29 04:40:16 crc kubenswrapper[4799]: I1129 04:40:16.972521 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 04:40:16 crc kubenswrapper[4799]: E1129 04:40:16.976908 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 04:40:17.476883296 +0000 UTC m=+93.119813686 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:17 crc kubenswrapper[4799]: I1129 04:40:17.037370 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-2qb64" podStartSLOduration=68.037342124 podStartE2EDuration="1m8.037342124s" podCreationTimestamp="2025-11-29 04:39:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:17.035039235 +0000 UTC m=+92.677969635" watchObservedRunningTime="2025-11-29 04:40:17.037342124 +0000 UTC m=+92.680272524" Nov 29 04:40:17 crc kubenswrapper[4799]: I1129 04:40:17.078196 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:17 crc kubenswrapper[4799]: E1129 04:40:17.078827 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 04:40:17.578815163 +0000 UTC m=+93.221745553 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2bbp8" (UID: "48c7bbb5-36e2-4677-b259-d3da6302d415") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:17 crc kubenswrapper[4799]: I1129 04:40:17.128585 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kp5gm" podStartSLOduration=68.128553935 podStartE2EDuration="1m8.128553935s" podCreationTimestamp="2025-11-29 04:39:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:17.06975232 +0000 UTC m=+92.712682720" watchObservedRunningTime="2025-11-29 04:40:17.128553935 +0000 UTC m=+92.771484335" Nov 29 04:40:17 crc kubenswrapper[4799]: I1129 04:40:17.130280 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6gm29" podStartSLOduration=68.13027355 podStartE2EDuration="1m8.13027355s" podCreationTimestamp="2025-11-29 04:39:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:17.130031484 +0000 UTC m=+92.772961884" watchObservedRunningTime="2025-11-29 04:40:17.13027355 +0000 UTC m=+92.773203950" Nov 29 04:40:17 crc kubenswrapper[4799]: I1129 04:40:17.184065 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 04:40:17 crc kubenswrapper[4799]: E1129 04:40:17.184366 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 04:40:17.684350834 +0000 UTC m=+93.327281234 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:17 crc kubenswrapper[4799]: I1129 04:40:17.220499 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-z7df5" podStartSLOduration=68.220468215 podStartE2EDuration="1m8.220468215s" podCreationTimestamp="2025-11-29 04:39:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:17.165428386 +0000 UTC m=+92.808358786" watchObservedRunningTime="2025-11-29 04:40:17.220468215 +0000 UTC m=+92.863398615" Nov 29 04:40:17 crc kubenswrapper[4799]: I1129 04:40:17.221396 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-ddl6h" podStartSLOduration=68.221391778 podStartE2EDuration="1m8.221391778s" podCreationTimestamp="2025-11-29 04:39:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:17.213739812 +0000 UTC m=+92.856670212" watchObservedRunningTime="2025-11-29 04:40:17.221391778 +0000 UTC m=+92.864322178" Nov 29 04:40:17 crc kubenswrapper[4799]: I1129 04:40:17.239014 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-85llw" podStartSLOduration=68.238999483 podStartE2EDuration="1m8.238999483s" podCreationTimestamp="2025-11-29 04:39:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:17.238168411 +0000 UTC m=+92.881098811" watchObservedRunningTime="2025-11-29 04:40:17.238999483 +0000 UTC m=+92.881929883" Nov 29 04:40:17 crc kubenswrapper[4799]: I1129 04:40:17.287554 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:17 crc kubenswrapper[4799]: E1129 04:40:17.287874 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 04:40:17.787863223 +0000 UTC m=+93.430793623 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2bbp8" (UID: "48c7bbb5-36e2-4677-b259-d3da6302d415") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:17 crc kubenswrapper[4799]: I1129 04:40:17.325968 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-nmvwb"] Nov 29 04:40:17 crc kubenswrapper[4799]: I1129 04:40:17.349492 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-5btb4" podStartSLOduration=68.349473841 podStartE2EDuration="1m8.349473841s" podCreationTimestamp="2025-11-29 04:39:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:17.347413497 +0000 UTC m=+92.990343947" watchObservedRunningTime="2025-11-29 04:40:17.349473841 +0000 UTC m=+92.992404241" Nov 29 04:40:17 crc kubenswrapper[4799]: I1129 04:40:17.357978 4799 patch_prober.go:28] interesting pod/router-default-5444994796-jrhwp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 29 04:40:17 crc kubenswrapper[4799]: [-]has-synced failed: reason withheld Nov 29 04:40:17 crc kubenswrapper[4799]: [+]process-running ok Nov 29 04:40:17 crc kubenswrapper[4799]: healthz check failed Nov 29 04:40:17 crc kubenswrapper[4799]: I1129 04:40:17.358030 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jrhwp" podUID="36a3c13c-2670-4b33-a9fe-9071b35365fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 29 04:40:17 crc kubenswrapper[4799]: I1129 04:40:17.379954 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4bfz8" podStartSLOduration=68.379934956 podStartE2EDuration="1m8.379934956s" podCreationTimestamp="2025-11-29 04:39:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:17.377897793 +0000 UTC m=+93.020828193" watchObservedRunningTime="2025-11-29 04:40:17.379934956 +0000 UTC m=+93.022865346" Nov 29 04:40:17 crc kubenswrapper[4799]: I1129 04:40:17.389624 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 04:40:17 crc kubenswrapper[4799]: E1129 04:40:17.389944 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 04:40:17.889929984 +0000 UTC m=+93.532860384 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:17 crc kubenswrapper[4799]: I1129 04:40:17.423472 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=1.4234522379999999 podStartE2EDuration="1.423452238s" podCreationTimestamp="2025-11-29 04:40:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:17.412499045 +0000 UTC m=+93.055429445" watchObservedRunningTime="2025-11-29 04:40:17.423452238 +0000 UTC m=+93.066382638" Nov 29 04:40:17 crc kubenswrapper[4799]: I1129 04:40:17.426729 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-jjnx9"] Nov 29 04:40:17 crc kubenswrapper[4799]: I1129 04:40:17.499231 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:17 crc kubenswrapper[4799]: E1129 04:40:17.499641 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 04:40:17.999629181 +0000 UTC m=+93.642559581 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2bbp8" (UID: "48c7bbb5-36e2-4677-b259-d3da6302d415") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:17 crc kubenswrapper[4799]: I1129 04:40:17.601495 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 04:40:17 crc kubenswrapper[4799]: E1129 04:40:17.601933 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 04:40:18.101913168 +0000 UTC m=+93.744843568 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:17 crc kubenswrapper[4799]: I1129 04:40:17.658613 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-vbffw"] Nov 29 04:40:17 crc kubenswrapper[4799]: I1129 04:40:17.704488 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:17 crc kubenswrapper[4799]: E1129 04:40:17.705108 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 04:40:18.205092888 +0000 UTC m=+93.848023288 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2bbp8" (UID: "48c7bbb5-36e2-4677-b259-d3da6302d415") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:17 crc kubenswrapper[4799]: I1129 04:40:17.805436 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 04:40:17 crc kubenswrapper[4799]: E1129 04:40:17.805565 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 04:40:18.305542777 +0000 UTC m=+93.948473177 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:17 crc kubenswrapper[4799]: I1129 04:40:17.805645 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:17 crc kubenswrapper[4799]: E1129 04:40:17.806998 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 04:40:18.306775359 +0000 UTC m=+93.949705749 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2bbp8" (UID: "48c7bbb5-36e2-4677-b259-d3da6302d415") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:17 crc kubenswrapper[4799]: I1129 04:40:17.915832 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 04:40:17 crc kubenswrapper[4799]: E1129 04:40:17.916135 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 04:40:18.416111628 +0000 UTC m=+94.059042028 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:17 crc kubenswrapper[4799]: I1129 04:40:17.929090 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:17 crc kubenswrapper[4799]: E1129 04:40:17.929723 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 04:40:18.429710638 +0000 UTC m=+94.072641038 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2bbp8" (UID: "48c7bbb5-36e2-4677-b259-d3da6302d415") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.002299 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t4f4r" event={"ID":"482e1948-ee9d-401d-a590-ed64818891d3","Type":"ContainerStarted","Data":"99d66558223f1bdb3f878acbe9bfec43b5c99b431f6ca63a24d6e8fbe261b4f4"} Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.030485 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 04:40:18 crc kubenswrapper[4799]: E1129 04:40:18.030808 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 04:40:18.530779524 +0000 UTC m=+94.173709914 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.061487 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-hkpdn" event={"ID":"9f2aa3db-bf5b-47b3-bcec-5132916f0f7d","Type":"ContainerStarted","Data":"16e3dfface9ea088e8cff0bbedd4576692e92dc04ad28d1a8602d102accf69b3"} Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.092999 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-rhzbg" event={"ID":"fecaea20-4e1a-4977-92e0-d0bd7c3037a2","Type":"ContainerStarted","Data":"24926035b65c4efde5d76e8b70deb5e7d7dbd2e35d5c73226ad419f1280bdae9"} Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.094137 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-rhzbg" Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.102345 4799 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-rhzbg container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.40:8080/healthz\": dial tcp 10.217.0.40:8080: connect: connection refused" start-of-body= Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.102398 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-rhzbg" podUID="fecaea20-4e1a-4977-92e0-d0bd7c3037a2" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.40:8080/healthz\": dial tcp 10.217.0.40:8080: connect: connection refused" Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.129130 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-tqr7m" event={"ID":"b22a0046-b9c4-4b6e-ba05-1c453c998ff8","Type":"ContainerStarted","Data":"4cf4c0916d97e5ddc335e26badfcbf93fe54e5c0058ed9718d8e2f40d3256243"} Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.130114 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-tqr7m" Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.131671 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:18 crc kubenswrapper[4799]: E1129 04:40:18.132059 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 04:40:18.632047495 +0000 UTC m=+94.274977895 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2bbp8" (UID: "48c7bbb5-36e2-4677-b259-d3da6302d415") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.179356 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9krqq" event={"ID":"3a945cd2-2261-435a-8d3f-bacd5b808d26","Type":"ContainerStarted","Data":"acfba574c21ac18d3ea99e42abc0fd21031465b070d6bc05dc1d0644cc7b8445"} Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.180376 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9krqq" Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.200273 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qz5br" event={"ID":"2f1d8a27-12bd-4ac8-b23f-51f9f10fbcbf","Type":"ContainerStarted","Data":"2ffb5034fcd5b77ec64130a8da5c85f39640bc9ec3ddcc9e160cd32eab4deeed"} Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.201102 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qz5br" Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.205755 4799 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-9krqq container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" start-of-body= Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.205816 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9krqq" podUID="3a945cd2-2261-435a-8d3f-bacd5b808d26" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.210693 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-bpxtd" event={"ID":"ee4ae924-84bb-4b5c-9b6e-f03d6cba2ad3","Type":"ContainerStarted","Data":"2b696edcd55fc3a48c7bfda26a693d5ba4bc21cb4c0a7e71bed27f38287ddf34"} Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.211513 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-hkpdn" podStartSLOduration=69.211502283 podStartE2EDuration="1m9.211502283s" podCreationTimestamp="2025-11-29 04:39:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:18.123536826 +0000 UTC m=+93.766467216" watchObservedRunningTime="2025-11-29 04:40:18.211502283 +0000 UTC m=+93.854432683" Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.213074 4799 generic.go:334] "Generic (PLEG): container finished" podID="34ecc5fd-acf1-4748-af5d-d9af38bff6d3" containerID="3c424e35796cbfcb8dd209151ec1346cee9758eee085362544bfa7b67cd6e7c6" exitCode=0 Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.213135 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-rcsd7" event={"ID":"34ecc5fd-acf1-4748-af5d-d9af38bff6d3","Type":"ContainerDied","Data":"3c424e35796cbfcb8dd209151ec1346cee9758eee085362544bfa7b67cd6e7c6"} Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.233221 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 04:40:18 crc kubenswrapper[4799]: E1129 04:40:18.234624 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 04:40:18.734608409 +0000 UTC m=+94.377538809 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.255181 4799 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-qz5br container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.18:8443/healthz\": dial tcp 10.217.0.18:8443: connect: connection refused" start-of-body= Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.255230 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qz5br" podUID="2f1d8a27-12bd-4ac8-b23f-51f9f10fbcbf" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.18:8443/healthz\": dial tcp 10.217.0.18:8443: connect: connection refused" Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.256295 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qd7d2" event={"ID":"cddf4720-7bab-43e0-989e-6cff2942a5fe","Type":"ContainerStarted","Data":"ec624215987266efc9b5108eb3c52d8923e7ca0e98e97de5e48bc90a74213a8e"} Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.256871 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qd7d2" Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.260049 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xj4jw" event={"ID":"74f1d4e8-2506-43c8-bcbb-cb8bf28cfb35","Type":"ContainerStarted","Data":"b24f5316b9a4db246b74a69d39f21afb31f62a788fc522620e63e594a24373de"} Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.280023 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6jsdb" event={"ID":"9f245325-f6b0-4ffa-b423-63a47df33136","Type":"ContainerStarted","Data":"12441574ce8dde9e077689427feab4a7323dbd274c28ffbed914e3ebe051d3f8"} Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.300562 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h4w54" event={"ID":"0f47a5c5-7c73-48ca-93a2-25d2973142d8","Type":"ContainerStarted","Data":"191e178153116877b1707fbed086dad649ee7885d2e3aa62fb3116ff5c1f4dd2"} Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.323231 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-6rp4p" event={"ID":"6cf3da23-866e-4182-bc35-021d2151f9f2","Type":"ContainerStarted","Data":"628cf03d3e3dd0f9a2f9abd55b130e94ff3d6b5b021d641ce1ed93191c2aa57a"} Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.334832 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.335303 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-rhzbg" podStartSLOduration=69.335287694 podStartE2EDuration="1m9.335287694s" podCreationTimestamp="2025-11-29 04:39:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:18.209936092 +0000 UTC m=+93.852866482" watchObservedRunningTime="2025-11-29 04:40:18.335287694 +0000 UTC m=+93.978218094" Nov 29 04:40:18 crc kubenswrapper[4799]: E1129 04:40:18.337706 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 04:40:18.837693997 +0000 UTC m=+94.480624397 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2bbp8" (UID: "48c7bbb5-36e2-4677-b259-d3da6302d415") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.338312 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qz5br" podStartSLOduration=69.338289672 podStartE2EDuration="1m9.338289672s" podCreationTimestamp="2025-11-29 04:39:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:18.322497655 +0000 UTC m=+93.965428055" watchObservedRunningTime="2025-11-29 04:40:18.338289672 +0000 UTC m=+93.981220072" Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.354930 4799 patch_prober.go:28] interesting pod/router-default-5444994796-jrhwp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 29 04:40:18 crc kubenswrapper[4799]: [-]has-synced failed: reason withheld Nov 29 04:40:18 crc kubenswrapper[4799]: [+]process-running ok Nov 29 04:40:18 crc kubenswrapper[4799]: healthz check failed Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.358021 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jrhwp" podUID="36a3c13c-2670-4b33-a9fe-9071b35365fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.389092 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-svdzq" event={"ID":"938dcbc0-ed30-4200-9f6f-cbc1364cce88","Type":"ContainerStarted","Data":"77314c57b93d59621654956bec87a9689cdac65a05c56322414791edc159111c"} Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.394042 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-svdzq" Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.421891 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-jjnx9" event={"ID":"f4a793d9-fe58-4613-bbd4-1550b95bd15a","Type":"ContainerStarted","Data":"888d91bb948d59d4541419e840044f6cfc7e164caca1806f405d46fc9efeb690"} Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.427992 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-2wtrf" event={"ID":"aef7387f-d081-4d53-bdb6-33f61918b318","Type":"ContainerStarted","Data":"4d365da74b2c9e445e6d10e5eabeae7b5e874aca0381a4bd9a3012ebc5129871"} Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.435638 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-tqr7m" podStartSLOduration=69.43560167 podStartE2EDuration="1m9.43560167s" podCreationTimestamp="2025-11-29 04:39:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:18.434924883 +0000 UTC m=+94.077855283" watchObservedRunningTime="2025-11-29 04:40:18.43560167 +0000 UTC m=+94.078532070" Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.436969 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 04:40:18 crc kubenswrapper[4799]: E1129 04:40:18.438545 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 04:40:18.938529196 +0000 UTC m=+94.581459596 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.518241 4799 generic.go:334] "Generic (PLEG): container finished" podID="191743db-39d2-4f93-add6-471de6ccdede" containerID="a78864dafae3451d75c5725c796a3a4dbd4da37178417b9e1d4fedf118437a22" exitCode=0 Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.518341 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zv6wk" event={"ID":"191743db-39d2-4f93-add6-471de6ccdede","Type":"ContainerDied","Data":"a78864dafae3451d75c5725c796a3a4dbd4da37178417b9e1d4fedf118437a22"} Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.538171 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.539321 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29406510-f2lz7" event={"ID":"96a1e39b-0018-41b8-8d7c-9fa7697858e9","Type":"ContainerStarted","Data":"4cdf55e51f971d2c8da1cd6d4fe65f3ad6e77e9db99c9d9f0bb9edbd026a4d63"} Nov 29 04:40:18 crc kubenswrapper[4799]: E1129 04:40:18.539555 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 04:40:19.03954578 +0000 UTC m=+94.682476180 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2bbp8" (UID: "48c7bbb5-36e2-4677-b259-d3da6302d415") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.580583 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-vbffw" event={"ID":"22463cfc-00a0-47f8-889a-35fea3fc4809","Type":"ContainerStarted","Data":"59778ec01fb977eb8b0109d7f33e635e846ab69aa4cfca100258b04cb5ffa854"} Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.617028 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" event={"ID":"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d","Type":"ContainerStarted","Data":"a6408959e5fbc68bed51a4dce9165a56d9c58cc0083289a8e606eb87e4511e8c"} Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.618094 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.630971 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9krqq" podStartSLOduration=69.630952436 podStartE2EDuration="1m9.630952436s" podCreationTimestamp="2025-11-29 04:39:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:18.536220024 +0000 UTC m=+94.179150424" watchObservedRunningTime="2025-11-29 04:40:18.630952436 +0000 UTC m=+94.273882836" Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.640262 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 04:40:18 crc kubenswrapper[4799]: E1129 04:40:18.641904 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 04:40:19.141886188 +0000 UTC m=+94.784816688 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.648184 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9j7n6" event={"ID":"19582088-032e-4b42-bc19-5080f7c97e13","Type":"ContainerStarted","Data":"59d5724e0d58bc7a72911cfe23c57810f4468e7b79f00801b7f530f844ce4a2f"} Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.724167 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-vh8lz" event={"ID":"2aac0475-5e98-4db7-a0f4-6f128acb77c0","Type":"ContainerStarted","Data":"d125f773b479bbf24ca318124b7074e712128e7a9db78dcc6e36df17797e0930"} Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.724201 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hs96x" event={"ID":"aab927aa-d4cb-4b96-a5a6-bb14df8d4af8","Type":"ContainerStarted","Data":"0d1a36d5df7ccd2d064d6990083d8d160db3d23356fc70f400c0fa6efde29c58"} Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.725264 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-6rp4p" podStartSLOduration=69.725250277 podStartE2EDuration="1m9.725250277s" podCreationTimestamp="2025-11-29 04:39:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:18.636579921 +0000 UTC m=+94.279510321" watchObservedRunningTime="2025-11-29 04:40:18.725250277 +0000 UTC m=+94.368180677" Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.725458 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xj4jw" podStartSLOduration=69.725452993 podStartE2EDuration="1m9.725452993s" podCreationTimestamp="2025-11-29 04:39:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:18.72497078 +0000 UTC m=+94.367901180" watchObservedRunningTime="2025-11-29 04:40:18.725452993 +0000 UTC m=+94.368383393" Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.742729 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:18 crc kubenswrapper[4799]: E1129 04:40:18.744978 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 04:40:19.244962805 +0000 UTC m=+94.887893275 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2bbp8" (UID: "48c7bbb5-36e2-4677-b259-d3da6302d415") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.759820 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jq22x" event={"ID":"f2baa9d1-df52-44c6-b62f-f39c16b70e34","Type":"ContainerStarted","Data":"39444df0592f2b402fbd73866ba082b4e4fb8deee0aafc1cce90b3bf2eac0566"} Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.775505 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-svdzq" podStartSLOduration=69.775481742 podStartE2EDuration="1m9.775481742s" podCreationTimestamp="2025-11-29 04:39:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:18.769453077 +0000 UTC m=+94.412383477" watchObservedRunningTime="2025-11-29 04:40:18.775481742 +0000 UTC m=+94.418412142" Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.821075 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g8b88" event={"ID":"6743e3ae-38d5-4ac1-8eea-7f0fbbe282da","Type":"ContainerStarted","Data":"6354dc3072b402bbce0e02c1aa254b3061b4beb3ad878403b10f13c06bdc5d16"} Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.822766 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g8b88" Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.845232 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 04:40:18 crc kubenswrapper[4799]: E1129 04:40:18.853267 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 04:40:19.353244087 +0000 UTC m=+94.996174487 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.892609 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-fnhgz" event={"ID":"50764326-a80d-4b7a-bfd3-2242f4d0e228","Type":"ContainerStarted","Data":"0efc6a167e755e40ef6a1eb617392e538d466bab8a02444048997c51bca4d15b"} Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.917161 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nmvwb" event={"ID":"c957ce17-f774-454a-a5e2-20f8ca7be2f0","Type":"ContainerStarted","Data":"d4e6cdd893126b26f58e8c94299c122642f38b376f3b205265c36907960bbb7c"} Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.917203 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nmvwb" event={"ID":"c957ce17-f774-454a-a5e2-20f8ca7be2f0","Type":"ContainerStarted","Data":"ae90f22cd2a960c339d12a6da9ef54c086a86333bb2dbe123b31a3131ad3ffd1"} Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.917771 4799 patch_prober.go:28] interesting pod/downloads-7954f5f757-bbm46 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.28:8080/\": dial tcp 10.217.0.28:8080: connect: connection refused" start-of-body= Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.917856 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-bbm46" podUID="fdba8af3-a255-470f-bffd-ced0a45e988d" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.28:8080/\": dial tcp 10.217.0.28:8080: connect: connection refused" Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.938835 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-jjnx9" podStartSLOduration=7.938814263 podStartE2EDuration="7.938814263s" podCreationTimestamp="2025-11-29 04:40:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:18.938376232 +0000 UTC m=+94.581306632" watchObservedRunningTime="2025-11-29 04:40:18.938814263 +0000 UTC m=+94.581744663" Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.939292 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-bpxtd" podStartSLOduration=69.939286025 podStartE2EDuration="1m9.939286025s" podCreationTimestamp="2025-11-29 04:39:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:18.895326602 +0000 UTC m=+94.538257002" watchObservedRunningTime="2025-11-29 04:40:18.939286025 +0000 UTC m=+94.582216425" Nov 29 04:40:18 crc kubenswrapper[4799]: I1129 04:40:18.953358 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:18 crc kubenswrapper[4799]: E1129 04:40:18.959450 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 04:40:19.459433695 +0000 UTC m=+95.102364095 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2bbp8" (UID: "48c7bbb5-36e2-4677-b259-d3da6302d415") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:19 crc kubenswrapper[4799]: I1129 04:40:19.054568 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qd7d2" podStartSLOduration=70.054545397 podStartE2EDuration="1m10.054545397s" podCreationTimestamp="2025-11-29 04:39:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:19.007111364 +0000 UTC m=+94.650041784" watchObservedRunningTime="2025-11-29 04:40:19.054545397 +0000 UTC m=+94.697475797" Nov 29 04:40:19 crc kubenswrapper[4799]: I1129 04:40:19.055379 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 04:40:19 crc kubenswrapper[4799]: E1129 04:40:19.055442 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 04:40:19.555427009 +0000 UTC m=+95.198357409 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:19 crc kubenswrapper[4799]: I1129 04:40:19.061277 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:19 crc kubenswrapper[4799]: E1129 04:40:19.061650 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 04:40:19.5616381 +0000 UTC m=+95.204568490 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2bbp8" (UID: "48c7bbb5-36e2-4677-b259-d3da6302d415") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:19 crc kubenswrapper[4799]: I1129 04:40:19.150524 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6jsdb" podStartSLOduration=70.15050823 podStartE2EDuration="1m10.15050823s" podCreationTimestamp="2025-11-29 04:39:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:19.066658438 +0000 UTC m=+94.709588838" watchObservedRunningTime="2025-11-29 04:40:19.15050823 +0000 UTC m=+94.793438630" Nov 29 04:40:19 crc kubenswrapper[4799]: I1129 04:40:19.169296 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 04:40:19 crc kubenswrapper[4799]: E1129 04:40:19.169756 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 04:40:19.669742256 +0000 UTC m=+95.312672656 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:19 crc kubenswrapper[4799]: I1129 04:40:19.207960 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-svdzq" Nov 29 04:40:19 crc kubenswrapper[4799]: I1129 04:40:19.227920 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" Nov 29 04:40:19 crc kubenswrapper[4799]: I1129 04:40:19.255730 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h4w54" podStartSLOduration=70.255713942 podStartE2EDuration="1m10.255713942s" podCreationTimestamp="2025-11-29 04:39:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:19.198684592 +0000 UTC m=+94.841615002" watchObservedRunningTime="2025-11-29 04:40:19.255713942 +0000 UTC m=+94.898644332" Nov 29 04:40:19 crc kubenswrapper[4799]: I1129 04:40:19.272553 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:19 crc kubenswrapper[4799]: E1129 04:40:19.273116 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 04:40:19.773102731 +0000 UTC m=+95.416033131 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2bbp8" (UID: "48c7bbb5-36e2-4677-b259-d3da6302d415") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:19 crc kubenswrapper[4799]: I1129 04:40:19.356769 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-2wtrf" podStartSLOduration=8.356749056 podStartE2EDuration="8.356749056s" podCreationTimestamp="2025-11-29 04:40:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:19.279993599 +0000 UTC m=+94.922923999" watchObservedRunningTime="2025-11-29 04:40:19.356749056 +0000 UTC m=+94.999679446" Nov 29 04:40:19 crc kubenswrapper[4799]: I1129 04:40:19.359300 4799 patch_prober.go:28] interesting pod/router-default-5444994796-jrhwp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 29 04:40:19 crc kubenswrapper[4799]: [-]has-synced failed: reason withheld Nov 29 04:40:19 crc kubenswrapper[4799]: [+]process-running ok Nov 29 04:40:19 crc kubenswrapper[4799]: healthz check failed Nov 29 04:40:19 crc kubenswrapper[4799]: I1129 04:40:19.359346 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jrhwp" podUID="36a3c13c-2670-4b33-a9fe-9071b35365fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 29 04:40:19 crc kubenswrapper[4799]: I1129 04:40:19.375638 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 04:40:19 crc kubenswrapper[4799]: E1129 04:40:19.376058 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 04:40:19.876043054 +0000 UTC m=+95.518973454 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:19 crc kubenswrapper[4799]: I1129 04:40:19.477704 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:19 crc kubenswrapper[4799]: E1129 04:40:19.478237 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 04:40:19.978223268 +0000 UTC m=+95.621153668 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2bbp8" (UID: "48c7bbb5-36e2-4677-b259-d3da6302d415") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:19 crc kubenswrapper[4799]: I1129 04:40:19.577505 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hs96x" podStartSLOduration=70.577486147 podStartE2EDuration="1m10.577486147s" podCreationTimestamp="2025-11-29 04:39:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:19.359714314 +0000 UTC m=+95.002644734" watchObservedRunningTime="2025-11-29 04:40:19.577486147 +0000 UTC m=+95.220416547" Nov 29 04:40:19 crc kubenswrapper[4799]: I1129 04:40:19.579364 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 04:40:19 crc kubenswrapper[4799]: E1129 04:40:19.579816 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 04:40:20.079767916 +0000 UTC m=+95.722698316 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:19 crc kubenswrapper[4799]: I1129 04:40:19.672559 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9j7n6" podStartSLOduration=70.672530047 podStartE2EDuration="1m10.672530047s" podCreationTimestamp="2025-11-29 04:39:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:19.575596669 +0000 UTC m=+95.218527069" watchObservedRunningTime="2025-11-29 04:40:19.672530047 +0000 UTC m=+95.315460447" Nov 29 04:40:19 crc kubenswrapper[4799]: I1129 04:40:19.680687 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:19 crc kubenswrapper[4799]: E1129 04:40:19.681139 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 04:40:20.181125409 +0000 UTC m=+95.824055809 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2bbp8" (UID: "48c7bbb5-36e2-4677-b259-d3da6302d415") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:19 crc kubenswrapper[4799]: I1129 04:40:19.781380 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 04:40:19 crc kubenswrapper[4799]: E1129 04:40:19.781783 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 04:40:20.281769094 +0000 UTC m=+95.924699494 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:19 crc kubenswrapper[4799]: I1129 04:40:19.822963 4799 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-g8b88 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.42:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 29 04:40:19 crc kubenswrapper[4799]: I1129 04:40:19.823025 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g8b88" podUID="6743e3ae-38d5-4ac1-8eea-7f0fbbe282da" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.42:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 29 04:40:19 crc kubenswrapper[4799]: I1129 04:40:19.863832 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" podStartSLOduration=70.863806409 podStartE2EDuration="1m10.863806409s" podCreationTimestamp="2025-11-29 04:39:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:19.862010902 +0000 UTC m=+95.504941302" watchObservedRunningTime="2025-11-29 04:40:19.863806409 +0000 UTC m=+95.506736809" Nov 29 04:40:19 crc kubenswrapper[4799]: I1129 04:40:19.866082 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-fnhgz" podStartSLOduration=70.866077057 podStartE2EDuration="1m10.866077057s" podCreationTimestamp="2025-11-29 04:39:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:19.678300876 +0000 UTC m=+95.321231276" watchObservedRunningTime="2025-11-29 04:40:19.866077057 +0000 UTC m=+95.509007457" Nov 29 04:40:19 crc kubenswrapper[4799]: I1129 04:40:19.882833 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:19 crc kubenswrapper[4799]: E1129 04:40:19.883318 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 04:40:20.38329721 +0000 UTC m=+96.026227610 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2bbp8" (UID: "48c7bbb5-36e2-4677-b259-d3da6302d415") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:19 crc kubenswrapper[4799]: I1129 04:40:19.941565 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9j7n6" event={"ID":"19582088-032e-4b42-bc19-5080f7c97e13","Type":"ContainerStarted","Data":"c81a1547c45d49117e77e0becf83796e448beff80a1bd969f47a9f6d4aec72fe"} Nov 29 04:40:19 crc kubenswrapper[4799]: I1129 04:40:19.949505 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h4w54" event={"ID":"0f47a5c5-7c73-48ca-93a2-25d2973142d8","Type":"ContainerStarted","Data":"1399e9f90f8aa6b4a13a3937963448ad6b3342c3ddf7a2fb890092e7585768b6"} Nov 29 04:40:19 crc kubenswrapper[4799]: I1129 04:40:19.953152 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-vbffw" event={"ID":"22463cfc-00a0-47f8-889a-35fea3fc4809","Type":"ContainerStarted","Data":"7b363764c4b2cba32c2eecd051e0710187f5fc363967307d9e05d929bcde50d8"} Nov 29 04:40:19 crc kubenswrapper[4799]: I1129 04:40:19.953179 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-vbffw" event={"ID":"22463cfc-00a0-47f8-889a-35fea3fc4809","Type":"ContainerStarted","Data":"b1900960b5ce70f6f24d2df20df25aaeac1a5b27086552ed9105a31579c71d78"} Nov 29 04:40:19 crc kubenswrapper[4799]: I1129 04:40:19.953587 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-vbffw" Nov 29 04:40:19 crc kubenswrapper[4799]: I1129 04:40:19.970111 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zv6wk" event={"ID":"191743db-39d2-4f93-add6-471de6ccdede","Type":"ContainerStarted","Data":"deb8a6f6b457113176c29e91ebb0ef4d8acd6946683e0e0d44aa8eeaf7256b68"} Nov 29 04:40:19 crc kubenswrapper[4799]: I1129 04:40:19.984582 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 04:40:19 crc kubenswrapper[4799]: E1129 04:40:19.986136 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 04:40:20.486119472 +0000 UTC m=+96.129049872 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:19 crc kubenswrapper[4799]: I1129 04:40:19.995441 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qd7d2" event={"ID":"cddf4720-7bab-43e0-989e-6cff2942a5fe","Type":"ContainerStarted","Data":"8feeea1ed700b21b0488f5782f9c1e94bb7282a30e899749ef5b38f85b682d44"} Nov 29 04:40:19 crc kubenswrapper[4799]: I1129 04:40:19.996445 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jq22x" podStartSLOduration=70.996433977 podStartE2EDuration="1m10.996433977s" podCreationTimestamp="2025-11-29 04:39:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:19.924667987 +0000 UTC m=+95.567598387" watchObservedRunningTime="2025-11-29 04:40:19.996433977 +0000 UTC m=+95.639364377" Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.021669 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-jjnx9" event={"ID":"f4a793d9-fe58-4613-bbd4-1550b95bd15a","Type":"ContainerStarted","Data":"05c83692e897d478f88cf66a7d93ab8dca9a3fb2085e06315b09c4d579a74a68"} Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.037890 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-rcsd7" event={"ID":"34ecc5fd-acf1-4748-af5d-d9af38bff6d3","Type":"ContainerStarted","Data":"a3b0d88f863ce7628c80683491e63a1483f1eecaaeaa1bd356f534b567a531d4"} Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.043734 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-rcsd7" event={"ID":"34ecc5fd-acf1-4748-af5d-d9af38bff6d3","Type":"ContainerStarted","Data":"f35a3cd07d747174c7e211a0e8f256107bde86f704ce5a8e03158fd364d4a281"} Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.067275 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nmvwb" podStartSLOduration=71.067250283 podStartE2EDuration="1m11.067250283s" podCreationTimestamp="2025-11-29 04:39:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:20.002939875 +0000 UTC m=+95.645870275" watchObservedRunningTime="2025-11-29 04:40:20.067250283 +0000 UTC m=+95.710180673" Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.086484 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-vh8lz" event={"ID":"2aac0475-5e98-4db7-a0f4-6f128acb77c0","Type":"ContainerStarted","Data":"5820f635e89fe227104a2b82aa95afb9e877952bc321d6e0596432534d61e88a"} Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.089715 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:20 crc kubenswrapper[4799]: E1129 04:40:20.099416 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 04:40:20.599396482 +0000 UTC m=+96.242326882 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2bbp8" (UID: "48c7bbb5-36e2-4677-b259-d3da6302d415") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.110717 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g8b88" podStartSLOduration=71.110693353 podStartE2EDuration="1m11.110693353s" podCreationTimestamp="2025-11-29 04:39:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:20.098357875 +0000 UTC m=+95.741288285" watchObservedRunningTime="2025-11-29 04:40:20.110693353 +0000 UTC m=+95.753623753" Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.134352 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nmvwb" event={"ID":"c957ce17-f774-454a-a5e2-20f8ca7be2f0","Type":"ContainerStarted","Data":"b0f1ba7d6c126302cba0dd9d266def28c94d85ad4dfa94a749cf44980eb195ed"} Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.147698 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29406510-f2lz7" podStartSLOduration=71.147677636 podStartE2EDuration="1m11.147677636s" podCreationTimestamp="2025-11-29 04:39:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:20.123697838 +0000 UTC m=+95.766628238" watchObservedRunningTime="2025-11-29 04:40:20.147677636 +0000 UTC m=+95.790608036" Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.172228 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-2t46v" event={"ID":"e69469d9-e3b1-4afd-84de-9b7dafa8e9e4","Type":"ContainerStarted","Data":"61c92a4ff1919fd79326ca290ba1e0a03aeb28f78189d8a6cfa9e4a1b8f916d9"} Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.172292 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-2t46v" event={"ID":"e69469d9-e3b1-4afd-84de-9b7dafa8e9e4","Type":"ContainerStarted","Data":"1d722a22b5f787e4e0d0efa7087249a985c7c70c9e028983d99a7b975dae2ced"} Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.198367 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 04:40:20 crc kubenswrapper[4799]: E1129 04:40:20.199939 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 04:40:20.699924124 +0000 UTC m=+96.342854524 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.211388 4799 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-rhzbg container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.40:8080/healthz\": dial tcp 10.217.0.40:8080: connect: connection refused" start-of-body= Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.211436 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-rhzbg" podUID="fecaea20-4e1a-4977-92e0-d0bd7c3037a2" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.40:8080/healthz\": dial tcp 10.217.0.40:8080: connect: connection refused" Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.211701 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t4f4r" event={"ID":"482e1948-ee9d-401d-a590-ed64818891d3","Type":"ContainerStarted","Data":"d1d9248c23468c5e611152a6af9f22c81e39ac8678dac9f28226c5620237330f"} Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.224498 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qz5br" Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.239097 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-vbffw" podStartSLOduration=9.239075333 podStartE2EDuration="9.239075333s" podCreationTimestamp="2025-11-29 04:40:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:20.235236874 +0000 UTC m=+95.878167284" watchObservedRunningTime="2025-11-29 04:40:20.239075333 +0000 UTC m=+95.882005733" Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.240816 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zv6wk" podStartSLOduration=71.240808877 podStartE2EDuration="1m11.240808877s" podCreationTimestamp="2025-11-29 04:39:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:20.184015683 +0000 UTC m=+95.826946083" watchObservedRunningTime="2025-11-29 04:40:20.240808877 +0000 UTC m=+95.883739277" Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.292903 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9krqq" Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.303614 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:20 crc kubenswrapper[4799]: E1129 04:40:20.310326 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 04:40:20.810313849 +0000 UTC m=+96.453244249 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2bbp8" (UID: "48c7bbb5-36e2-4677-b259-d3da6302d415") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.319860 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t4f4r" podStartSLOduration=71.319843244 podStartE2EDuration="1m11.319843244s" podCreationTimestamp="2025-11-29 04:39:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:20.319109886 +0000 UTC m=+95.962040286" watchObservedRunningTime="2025-11-29 04:40:20.319843244 +0000 UTC m=+95.962773644" Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.322742 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-rcsd7" podStartSLOduration=71.322729219 podStartE2EDuration="1m11.322729219s" podCreationTimestamp="2025-11-29 04:39:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:20.281140227 +0000 UTC m=+95.924070627" watchObservedRunningTime="2025-11-29 04:40:20.322729219 +0000 UTC m=+95.965659629" Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.350947 4799 patch_prober.go:28] interesting pod/router-default-5444994796-jrhwp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 29 04:40:20 crc kubenswrapper[4799]: [-]has-synced failed: reason withheld Nov 29 04:40:20 crc kubenswrapper[4799]: [+]process-running ok Nov 29 04:40:20 crc kubenswrapper[4799]: healthz check failed Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.351256 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jrhwp" podUID="36a3c13c-2670-4b33-a9fe-9071b35365fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.405740 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 04:40:20 crc kubenswrapper[4799]: E1129 04:40:20.406419 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 04:40:20.906402696 +0000 UTC m=+96.549333086 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.412851 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-2t46v" podStartSLOduration=71.412836552 podStartE2EDuration="1m11.412836552s" podCreationTimestamp="2025-11-29 04:39:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:20.363350796 +0000 UTC m=+96.006281196" watchObservedRunningTime="2025-11-29 04:40:20.412836552 +0000 UTC m=+96.055766952" Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.508754 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:20 crc kubenswrapper[4799]: E1129 04:40:20.509144 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 04:40:21.009132664 +0000 UTC m=+96.652063064 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2bbp8" (UID: "48c7bbb5-36e2-4677-b259-d3da6302d415") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.610067 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 04:40:20 crc kubenswrapper[4799]: E1129 04:40:20.610522 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 04:40:21.110505118 +0000 UTC m=+96.753435508 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.644854 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g8b88" Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.711503 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:20 crc kubenswrapper[4799]: E1129 04:40:20.712239 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 04:40:21.212221939 +0000 UTC m=+96.855152339 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2bbp8" (UID: "48c7bbb5-36e2-4677-b259-d3da6302d415") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.730839 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jlvpn"] Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.731780 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jlvpn" Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.745551 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jlvpn"] Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.746828 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.813244 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 04:40:20 crc kubenswrapper[4799]: E1129 04:40:20.813442 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 04:40:21.313416109 +0000 UTC m=+96.956346509 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.813778 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6ssg\" (UniqueName: \"kubernetes.io/projected/2544bf9f-2bb8-4661-997c-9b25a96103d4-kube-api-access-n6ssg\") pod \"community-operators-jlvpn\" (UID: \"2544bf9f-2bb8-4661-997c-9b25a96103d4\") " pod="openshift-marketplace/community-operators-jlvpn" Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.813862 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2544bf9f-2bb8-4661-997c-9b25a96103d4-utilities\") pod \"community-operators-jlvpn\" (UID: \"2544bf9f-2bb8-4661-997c-9b25a96103d4\") " pod="openshift-marketplace/community-operators-jlvpn" Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.813921 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.813978 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2544bf9f-2bb8-4661-997c-9b25a96103d4-catalog-content\") pod \"community-operators-jlvpn\" (UID: \"2544bf9f-2bb8-4661-997c-9b25a96103d4\") " pod="openshift-marketplace/community-operators-jlvpn" Nov 29 04:40:20 crc kubenswrapper[4799]: E1129 04:40:20.814204 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 04:40:21.314195698 +0000 UTC m=+96.957126098 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2bbp8" (UID: "48c7bbb5-36e2-4677-b259-d3da6302d415") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.853563 4799 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.926347 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 04:40:20 crc kubenswrapper[4799]: E1129 04:40:20.926509 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 04:40:21.426486024 +0000 UTC m=+97.069416424 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.926584 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2544bf9f-2bb8-4661-997c-9b25a96103d4-utilities\") pod \"community-operators-jlvpn\" (UID: \"2544bf9f-2bb8-4661-997c-9b25a96103d4\") " pod="openshift-marketplace/community-operators-jlvpn" Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.926645 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.926677 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2544bf9f-2bb8-4661-997c-9b25a96103d4-catalog-content\") pod \"community-operators-jlvpn\" (UID: \"2544bf9f-2bb8-4661-997c-9b25a96103d4\") " pod="openshift-marketplace/community-operators-jlvpn" Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.926727 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6ssg\" (UniqueName: \"kubernetes.io/projected/2544bf9f-2bb8-4661-997c-9b25a96103d4-kube-api-access-n6ssg\") pod \"community-operators-jlvpn\" (UID: \"2544bf9f-2bb8-4661-997c-9b25a96103d4\") " pod="openshift-marketplace/community-operators-jlvpn" Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.927023 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2544bf9f-2bb8-4661-997c-9b25a96103d4-utilities\") pod \"community-operators-jlvpn\" (UID: \"2544bf9f-2bb8-4661-997c-9b25a96103d4\") " pod="openshift-marketplace/community-operators-jlvpn" Nov 29 04:40:20 crc kubenswrapper[4799]: E1129 04:40:20.927043 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 04:40:21.427022257 +0000 UTC m=+97.069952657 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2bbp8" (UID: "48c7bbb5-36e2-4677-b259-d3da6302d415") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.927319 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2544bf9f-2bb8-4661-997c-9b25a96103d4-catalog-content\") pod \"community-operators-jlvpn\" (UID: \"2544bf9f-2bb8-4661-997c-9b25a96103d4\") " pod="openshift-marketplace/community-operators-jlvpn" Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.936046 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-tqr7m" Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.957372 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dnz56"] Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.958628 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dnz56" Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.960708 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 29 04:40:20 crc kubenswrapper[4799]: I1129 04:40:20.975256 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dnz56"] Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.003723 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6ssg\" (UniqueName: \"kubernetes.io/projected/2544bf9f-2bb8-4661-997c-9b25a96103d4-kube-api-access-n6ssg\") pod \"community-operators-jlvpn\" (UID: \"2544bf9f-2bb8-4661-997c-9b25a96103d4\") " pod="openshift-marketplace/community-operators-jlvpn" Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.035856 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.036300 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8734a259-eaa8-459e-b38d-e821e336a863-catalog-content\") pod \"certified-operators-dnz56\" (UID: \"8734a259-eaa8-459e-b38d-e821e336a863\") " pod="openshift-marketplace/certified-operators-dnz56" Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.036393 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8734a259-eaa8-459e-b38d-e821e336a863-utilities\") pod \"certified-operators-dnz56\" (UID: \"8734a259-eaa8-459e-b38d-e821e336a863\") " pod="openshift-marketplace/certified-operators-dnz56" Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.036453 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b78n8\" (UniqueName: \"kubernetes.io/projected/8734a259-eaa8-459e-b38d-e821e336a863-kube-api-access-b78n8\") pod \"certified-operators-dnz56\" (UID: \"8734a259-eaa8-459e-b38d-e821e336a863\") " pod="openshift-marketplace/certified-operators-dnz56" Nov 29 04:40:21 crc kubenswrapper[4799]: E1129 04:40:21.036589 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 04:40:21.536572451 +0000 UTC m=+97.179502851 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.046041 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jlvpn" Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.145198 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b78n8\" (UniqueName: \"kubernetes.io/projected/8734a259-eaa8-459e-b38d-e821e336a863-kube-api-access-b78n8\") pod \"certified-operators-dnz56\" (UID: \"8734a259-eaa8-459e-b38d-e821e336a863\") " pod="openshift-marketplace/certified-operators-dnz56" Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.145535 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8734a259-eaa8-459e-b38d-e821e336a863-catalog-content\") pod \"certified-operators-dnz56\" (UID: \"8734a259-eaa8-459e-b38d-e821e336a863\") " pod="openshift-marketplace/certified-operators-dnz56" Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.145572 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.145601 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8734a259-eaa8-459e-b38d-e821e336a863-utilities\") pod \"certified-operators-dnz56\" (UID: \"8734a259-eaa8-459e-b38d-e821e336a863\") " pod="openshift-marketplace/certified-operators-dnz56" Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.145998 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8734a259-eaa8-459e-b38d-e821e336a863-catalog-content\") pod \"certified-operators-dnz56\" (UID: \"8734a259-eaa8-459e-b38d-e821e336a863\") " pod="openshift-marketplace/certified-operators-dnz56" Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.146027 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8734a259-eaa8-459e-b38d-e821e336a863-utilities\") pod \"certified-operators-dnz56\" (UID: \"8734a259-eaa8-459e-b38d-e821e336a863\") " pod="openshift-marketplace/certified-operators-dnz56" Nov 29 04:40:21 crc kubenswrapper[4799]: E1129 04:40:21.146261 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 04:40:21.646247608 +0000 UTC m=+97.289178008 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2bbp8" (UID: "48c7bbb5-36e2-4677-b259-d3da6302d415") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.196831 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mplj2"] Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.198242 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mplj2" Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.207137 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b78n8\" (UniqueName: \"kubernetes.io/projected/8734a259-eaa8-459e-b38d-e821e336a863-kube-api-access-b78n8\") pod \"certified-operators-dnz56\" (UID: \"8734a259-eaa8-459e-b38d-e821e336a863\") " pod="openshift-marketplace/certified-operators-dnz56" Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.218394 4799 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-11-29T04:40:20.853589744Z","Handler":null,"Name":""} Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.233711 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mplj2"] Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.241052 4799 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.241092 4799 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.242046 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-vh8lz" event={"ID":"2aac0475-5e98-4db7-a0f4-6f128acb77c0","Type":"ContainerStarted","Data":"6af4858e2b330bfa4d0f61de806bbb5228d53f07e1928c4e11626ccf1dfff6f1"} Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.242103 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-vh8lz" event={"ID":"2aac0475-5e98-4db7-a0f4-6f128acb77c0","Type":"ContainerStarted","Data":"8f8f9a4650de9176b60a24561e7b5ce469c75d401d7d823e03ce3e3f63d0c1e7"} Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.246217 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.252268 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-rhzbg" Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.282852 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dnz56" Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.284217 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.336921 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vl566"] Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.337868 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vl566" Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.351070 4799 patch_prober.go:28] interesting pod/router-default-5444994796-jrhwp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 29 04:40:21 crc kubenswrapper[4799]: [-]has-synced failed: reason withheld Nov 29 04:40:21 crc kubenswrapper[4799]: [+]process-running ok Nov 29 04:40:21 crc kubenswrapper[4799]: healthz check failed Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.351561 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jrhwp" podUID="36a3c13c-2670-4b33-a9fe-9071b35365fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.352769 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xvpw\" (UniqueName: \"kubernetes.io/projected/8a2cc255-bc5f-4253-bc2b-e74817971afe-kube-api-access-7xvpw\") pod \"community-operators-mplj2\" (UID: \"8a2cc255-bc5f-4253-bc2b-e74817971afe\") " pod="openshift-marketplace/community-operators-mplj2" Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.352895 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a2cc255-bc5f-4253-bc2b-e74817971afe-utilities\") pod \"community-operators-mplj2\" (UID: \"8a2cc255-bc5f-4253-bc2b-e74817971afe\") " pod="openshift-marketplace/community-operators-mplj2" Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.353071 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.353195 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a2cc255-bc5f-4253-bc2b-e74817971afe-catalog-content\") pod \"community-operators-mplj2\" (UID: \"8a2cc255-bc5f-4253-bc2b-e74817971afe\") " pod="openshift-marketplace/community-operators-mplj2" Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.363824 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vl566"] Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.456933 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xvpw\" (UniqueName: \"kubernetes.io/projected/8a2cc255-bc5f-4253-bc2b-e74817971afe-kube-api-access-7xvpw\") pod \"community-operators-mplj2\" (UID: \"8a2cc255-bc5f-4253-bc2b-e74817971afe\") " pod="openshift-marketplace/community-operators-mplj2" Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.457191 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a2cc255-bc5f-4253-bc2b-e74817971afe-utilities\") pod \"community-operators-mplj2\" (UID: \"8a2cc255-bc5f-4253-bc2b-e74817971afe\") " pod="openshift-marketplace/community-operators-mplj2" Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.457238 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d676x\" (UniqueName: \"kubernetes.io/projected/b0bccaa4-e99f-47b1-bf20-2895f289ffd7-kube-api-access-d676x\") pod \"certified-operators-vl566\" (UID: \"b0bccaa4-e99f-47b1-bf20-2895f289ffd7\") " pod="openshift-marketplace/certified-operators-vl566" Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.457305 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a2cc255-bc5f-4253-bc2b-e74817971afe-catalog-content\") pod \"community-operators-mplj2\" (UID: \"8a2cc255-bc5f-4253-bc2b-e74817971afe\") " pod="openshift-marketplace/community-operators-mplj2" Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.457324 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0bccaa4-e99f-47b1-bf20-2895f289ffd7-catalog-content\") pod \"certified-operators-vl566\" (UID: \"b0bccaa4-e99f-47b1-bf20-2895f289ffd7\") " pod="openshift-marketplace/certified-operators-vl566" Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.457353 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0bccaa4-e99f-47b1-bf20-2895f289ffd7-utilities\") pod \"certified-operators-vl566\" (UID: \"b0bccaa4-e99f-47b1-bf20-2895f289ffd7\") " pod="openshift-marketplace/certified-operators-vl566" Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.457684 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a2cc255-bc5f-4253-bc2b-e74817971afe-utilities\") pod \"community-operators-mplj2\" (UID: \"8a2cc255-bc5f-4253-bc2b-e74817971afe\") " pod="openshift-marketplace/community-operators-mplj2" Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.458108 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a2cc255-bc5f-4253-bc2b-e74817971afe-catalog-content\") pod \"community-operators-mplj2\" (UID: \"8a2cc255-bc5f-4253-bc2b-e74817971afe\") " pod="openshift-marketplace/community-operators-mplj2" Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.483216 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xvpw\" (UniqueName: \"kubernetes.io/projected/8a2cc255-bc5f-4253-bc2b-e74817971afe-kube-api-access-7xvpw\") pod \"community-operators-mplj2\" (UID: \"8a2cc255-bc5f-4253-bc2b-e74817971afe\") " pod="openshift-marketplace/community-operators-mplj2" Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.534125 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jlvpn"] Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.539732 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mplj2" Nov 29 04:40:21 crc kubenswrapper[4799]: W1129 04:40:21.549406 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2544bf9f_2bb8_4661_997c_9b25a96103d4.slice/crio-e02b6a5cc9c9ab9231d928449e29cf566f7b55db8a447c83cf25edae5434977b WatchSource:0}: Error finding container e02b6a5cc9c9ab9231d928449e29cf566f7b55db8a447c83cf25edae5434977b: Status 404 returned error can't find the container with id e02b6a5cc9c9ab9231d928449e29cf566f7b55db8a447c83cf25edae5434977b Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.559734 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d676x\" (UniqueName: \"kubernetes.io/projected/b0bccaa4-e99f-47b1-bf20-2895f289ffd7-kube-api-access-d676x\") pod \"certified-operators-vl566\" (UID: \"b0bccaa4-e99f-47b1-bf20-2895f289ffd7\") " pod="openshift-marketplace/certified-operators-vl566" Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.559831 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0bccaa4-e99f-47b1-bf20-2895f289ffd7-catalog-content\") pod \"certified-operators-vl566\" (UID: \"b0bccaa4-e99f-47b1-bf20-2895f289ffd7\") " pod="openshift-marketplace/certified-operators-vl566" Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.559865 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0bccaa4-e99f-47b1-bf20-2895f289ffd7-utilities\") pod \"certified-operators-vl566\" (UID: \"b0bccaa4-e99f-47b1-bf20-2895f289ffd7\") " pod="openshift-marketplace/certified-operators-vl566" Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.559909 4799 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.559983 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.560377 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0bccaa4-e99f-47b1-bf20-2895f289ffd7-utilities\") pod \"certified-operators-vl566\" (UID: \"b0bccaa4-e99f-47b1-bf20-2895f289ffd7\") " pod="openshift-marketplace/certified-operators-vl566" Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.561419 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0bccaa4-e99f-47b1-bf20-2895f289ffd7-catalog-content\") pod \"certified-operators-vl566\" (UID: \"b0bccaa4-e99f-47b1-bf20-2895f289ffd7\") " pod="openshift-marketplace/certified-operators-vl566" Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.586253 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d676x\" (UniqueName: \"kubernetes.io/projected/b0bccaa4-e99f-47b1-bf20-2895f289ffd7-kube-api-access-d676x\") pod \"certified-operators-vl566\" (UID: \"b0bccaa4-e99f-47b1-bf20-2895f289ffd7\") " pod="openshift-marketplace/certified-operators-vl566" Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.607463 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dnz56"] Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.659734 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2bbp8\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.721841 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vl566" Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.811742 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mplj2"] Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.918307 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:21 crc kubenswrapper[4799]: I1129 04:40:21.966399 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vl566"] Nov 29 04:40:21 crc kubenswrapper[4799]: W1129 04:40:21.971885 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb0bccaa4_e99f_47b1_bf20_2895f289ffd7.slice/crio-f6237c5e1897b877aa2f5608a2b48444b4d810c1670ad3a7cd116775cf533bfe WatchSource:0}: Error finding container f6237c5e1897b877aa2f5608a2b48444b4d810c1670ad3a7cd116775cf533bfe: Status 404 returned error can't find the container with id f6237c5e1897b877aa2f5608a2b48444b4d810c1670ad3a7cd116775cf533bfe Nov 29 04:40:22 crc kubenswrapper[4799]: I1129 04:40:22.148869 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-2bbp8"] Nov 29 04:40:22 crc kubenswrapper[4799]: W1129 04:40:22.154445 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod48c7bbb5_36e2_4677_b259_d3da6302d415.slice/crio-50327e8968746366dd798385527fba10ff6ff8d92c9618a20547417e602a4d54 WatchSource:0}: Error finding container 50327e8968746366dd798385527fba10ff6ff8d92c9618a20547417e602a4d54: Status 404 returned error can't find the container with id 50327e8968746366dd798385527fba10ff6ff8d92c9618a20547417e602a4d54 Nov 29 04:40:22 crc kubenswrapper[4799]: I1129 04:40:22.250839 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-vh8lz" event={"ID":"2aac0475-5e98-4db7-a0f4-6f128acb77c0","Type":"ContainerStarted","Data":"9b43ca91bef3538ed13bd5ae4550629bf1c105d86fd00d3ba0ebf34aeae075c2"} Nov 29 04:40:22 crc kubenswrapper[4799]: I1129 04:40:22.252459 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" event={"ID":"48c7bbb5-36e2-4677-b259-d3da6302d415","Type":"ContainerStarted","Data":"50327e8968746366dd798385527fba10ff6ff8d92c9618a20547417e602a4d54"} Nov 29 04:40:22 crc kubenswrapper[4799]: I1129 04:40:22.255049 4799 generic.go:334] "Generic (PLEG): container finished" podID="8a2cc255-bc5f-4253-bc2b-e74817971afe" containerID="c4cf32469a3d4feffd40673f38d4a39abb101d5196c09641054e3eff516aa22e" exitCode=0 Nov 29 04:40:22 crc kubenswrapper[4799]: I1129 04:40:22.255423 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mplj2" event={"ID":"8a2cc255-bc5f-4253-bc2b-e74817971afe","Type":"ContainerDied","Data":"c4cf32469a3d4feffd40673f38d4a39abb101d5196c09641054e3eff516aa22e"} Nov 29 04:40:22 crc kubenswrapper[4799]: I1129 04:40:22.255446 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mplj2" event={"ID":"8a2cc255-bc5f-4253-bc2b-e74817971afe","Type":"ContainerStarted","Data":"e25cd57ab8d2aa6363166cd416b592c66ebd00c0e1f196b0aeee28f1b480b110"} Nov 29 04:40:22 crc kubenswrapper[4799]: I1129 04:40:22.256399 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 29 04:40:22 crc kubenswrapper[4799]: I1129 04:40:22.257678 4799 generic.go:334] "Generic (PLEG): container finished" podID="b0bccaa4-e99f-47b1-bf20-2895f289ffd7" containerID="ffa7c3730d637df4273cd6ee511ae0064adcaee9e43dfafc39d7156a78f96706" exitCode=0 Nov 29 04:40:22 crc kubenswrapper[4799]: I1129 04:40:22.257742 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vl566" event={"ID":"b0bccaa4-e99f-47b1-bf20-2895f289ffd7","Type":"ContainerDied","Data":"ffa7c3730d637df4273cd6ee511ae0064adcaee9e43dfafc39d7156a78f96706"} Nov 29 04:40:22 crc kubenswrapper[4799]: I1129 04:40:22.257762 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vl566" event={"ID":"b0bccaa4-e99f-47b1-bf20-2895f289ffd7","Type":"ContainerStarted","Data":"f6237c5e1897b877aa2f5608a2b48444b4d810c1670ad3a7cd116775cf533bfe"} Nov 29 04:40:22 crc kubenswrapper[4799]: I1129 04:40:22.259558 4799 generic.go:334] "Generic (PLEG): container finished" podID="2544bf9f-2bb8-4661-997c-9b25a96103d4" containerID="0915c197c63cf6faf6c69d1fa7d87570a8cc2cf97b65a0b4a4e8b99d67ec95d3" exitCode=0 Nov 29 04:40:22 crc kubenswrapper[4799]: I1129 04:40:22.259655 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jlvpn" event={"ID":"2544bf9f-2bb8-4661-997c-9b25a96103d4","Type":"ContainerDied","Data":"0915c197c63cf6faf6c69d1fa7d87570a8cc2cf97b65a0b4a4e8b99d67ec95d3"} Nov 29 04:40:22 crc kubenswrapper[4799]: I1129 04:40:22.259702 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jlvpn" event={"ID":"2544bf9f-2bb8-4661-997c-9b25a96103d4","Type":"ContainerStarted","Data":"e02b6a5cc9c9ab9231d928449e29cf566f7b55db8a447c83cf25edae5434977b"} Nov 29 04:40:22 crc kubenswrapper[4799]: I1129 04:40:22.265323 4799 generic.go:334] "Generic (PLEG): container finished" podID="96a1e39b-0018-41b8-8d7c-9fa7697858e9" containerID="4cdf55e51f971d2c8da1cd6d4fe65f3ad6e77e9db99c9d9f0bb9edbd026a4d63" exitCode=0 Nov 29 04:40:22 crc kubenswrapper[4799]: I1129 04:40:22.265352 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29406510-f2lz7" event={"ID":"96a1e39b-0018-41b8-8d7c-9fa7697858e9","Type":"ContainerDied","Data":"4cdf55e51f971d2c8da1cd6d4fe65f3ad6e77e9db99c9d9f0bb9edbd026a4d63"} Nov 29 04:40:22 crc kubenswrapper[4799]: I1129 04:40:22.277752 4799 generic.go:334] "Generic (PLEG): container finished" podID="8734a259-eaa8-459e-b38d-e821e336a863" containerID="8eeb2ca354b67247aa5a76641a69ba54701ee09f72d0a4f3ceaac0ac1e7903d1" exitCode=0 Nov 29 04:40:22 crc kubenswrapper[4799]: I1129 04:40:22.277938 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-vh8lz" podStartSLOduration=11.277917903 podStartE2EDuration="11.277917903s" podCreationTimestamp="2025-11-29 04:40:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:22.276543577 +0000 UTC m=+97.919473977" watchObservedRunningTime="2025-11-29 04:40:22.277917903 +0000 UTC m=+97.920848303" Nov 29 04:40:22 crc kubenswrapper[4799]: I1129 04:40:22.279022 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dnz56" event={"ID":"8734a259-eaa8-459e-b38d-e821e336a863","Type":"ContainerDied","Data":"8eeb2ca354b67247aa5a76641a69ba54701ee09f72d0a4f3ceaac0ac1e7903d1"} Nov 29 04:40:22 crc kubenswrapper[4799]: I1129 04:40:22.279067 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dnz56" event={"ID":"8734a259-eaa8-459e-b38d-e821e336a863","Type":"ContainerStarted","Data":"5d6d52963f632c487a32a238c4563ce502479a251bf1b92c8118cb8184aec24d"} Nov 29 04:40:22 crc kubenswrapper[4799]: I1129 04:40:22.345566 4799 patch_prober.go:28] interesting pod/router-default-5444994796-jrhwp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 29 04:40:22 crc kubenswrapper[4799]: [-]has-synced failed: reason withheld Nov 29 04:40:22 crc kubenswrapper[4799]: [+]process-running ok Nov 29 04:40:22 crc kubenswrapper[4799]: healthz check failed Nov 29 04:40:22 crc kubenswrapper[4799]: I1129 04:40:22.345633 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jrhwp" podUID="36a3c13c-2670-4b33-a9fe-9071b35365fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 29 04:40:22 crc kubenswrapper[4799]: I1129 04:40:22.582314 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 29 04:40:22 crc kubenswrapper[4799]: I1129 04:40:22.583156 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 29 04:40:22 crc kubenswrapper[4799]: I1129 04:40:22.585401 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Nov 29 04:40:22 crc kubenswrapper[4799]: I1129 04:40:22.585415 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Nov 29 04:40:22 crc kubenswrapper[4799]: I1129 04:40:22.593394 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 29 04:40:22 crc kubenswrapper[4799]: I1129 04:40:22.671146 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Nov 29 04:40:22 crc kubenswrapper[4799]: I1129 04:40:22.702449 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2a4dbe70-2ff5-41cb-8d0a-02e4451fca96-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"2a4dbe70-2ff5-41cb-8d0a-02e4451fca96\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 29 04:40:22 crc kubenswrapper[4799]: I1129 04:40:22.702627 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2a4dbe70-2ff5-41cb-8d0a-02e4451fca96-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"2a4dbe70-2ff5-41cb-8d0a-02e4451fca96\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 29 04:40:22 crc kubenswrapper[4799]: I1129 04:40:22.804009 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2a4dbe70-2ff5-41cb-8d0a-02e4451fca96-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"2a4dbe70-2ff5-41cb-8d0a-02e4451fca96\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 29 04:40:22 crc kubenswrapper[4799]: I1129 04:40:22.804117 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2a4dbe70-2ff5-41cb-8d0a-02e4451fca96-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"2a4dbe70-2ff5-41cb-8d0a-02e4451fca96\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 29 04:40:22 crc kubenswrapper[4799]: I1129 04:40:22.804117 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2a4dbe70-2ff5-41cb-8d0a-02e4451fca96-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"2a4dbe70-2ff5-41cb-8d0a-02e4451fca96\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 29 04:40:22 crc kubenswrapper[4799]: I1129 04:40:22.827603 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2a4dbe70-2ff5-41cb-8d0a-02e4451fca96-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"2a4dbe70-2ff5-41cb-8d0a-02e4451fca96\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 29 04:40:22 crc kubenswrapper[4799]: I1129 04:40:22.854295 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 29 04:40:22 crc kubenswrapper[4799]: I1129 04:40:22.855038 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 29 04:40:22 crc kubenswrapper[4799]: I1129 04:40:22.857168 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 29 04:40:22 crc kubenswrapper[4799]: I1129 04:40:22.861975 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 29 04:40:22 crc kubenswrapper[4799]: I1129 04:40:22.862745 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 29 04:40:22 crc kubenswrapper[4799]: I1129 04:40:22.905275 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 29 04:40:22 crc kubenswrapper[4799]: I1129 04:40:22.914910 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-q6m25"] Nov 29 04:40:22 crc kubenswrapper[4799]: I1129 04:40:22.916122 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q6m25" Nov 29 04:40:22 crc kubenswrapper[4799]: I1129 04:40:22.926469 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 29 04:40:22 crc kubenswrapper[4799]: I1129 04:40:22.934773 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q6m25"] Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.006714 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8p5ww\" (UniqueName: \"kubernetes.io/projected/0a4df431-deee-45d6-9c07-7a1616a23d91-kube-api-access-8p5ww\") pod \"redhat-marketplace-q6m25\" (UID: \"0a4df431-deee-45d6-9c07-7a1616a23d91\") " pod="openshift-marketplace/redhat-marketplace-q6m25" Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.006775 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6efe6a93-4ea8-4383-b104-988fc603d3ef-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"6efe6a93-4ea8-4383-b104-988fc603d3ef\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.006908 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6efe6a93-4ea8-4383-b104-988fc603d3ef-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"6efe6a93-4ea8-4383-b104-988fc603d3ef\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.007131 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a4df431-deee-45d6-9c07-7a1616a23d91-catalog-content\") pod \"redhat-marketplace-q6m25\" (UID: \"0a4df431-deee-45d6-9c07-7a1616a23d91\") " pod="openshift-marketplace/redhat-marketplace-q6m25" Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.007351 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a4df431-deee-45d6-9c07-7a1616a23d91-utilities\") pod \"redhat-marketplace-q6m25\" (UID: \"0a4df431-deee-45d6-9c07-7a1616a23d91\") " pod="openshift-marketplace/redhat-marketplace-q6m25" Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.128444 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a4df431-deee-45d6-9c07-7a1616a23d91-utilities\") pod \"redhat-marketplace-q6m25\" (UID: \"0a4df431-deee-45d6-9c07-7a1616a23d91\") " pod="openshift-marketplace/redhat-marketplace-q6m25" Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.128671 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8p5ww\" (UniqueName: \"kubernetes.io/projected/0a4df431-deee-45d6-9c07-7a1616a23d91-kube-api-access-8p5ww\") pod \"redhat-marketplace-q6m25\" (UID: \"0a4df431-deee-45d6-9c07-7a1616a23d91\") " pod="openshift-marketplace/redhat-marketplace-q6m25" Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.128710 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6efe6a93-4ea8-4383-b104-988fc603d3ef-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"6efe6a93-4ea8-4383-b104-988fc603d3ef\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.128764 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6efe6a93-4ea8-4383-b104-988fc603d3ef-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"6efe6a93-4ea8-4383-b104-988fc603d3ef\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.128803 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a4df431-deee-45d6-9c07-7a1616a23d91-catalog-content\") pod \"redhat-marketplace-q6m25\" (UID: \"0a4df431-deee-45d6-9c07-7a1616a23d91\") " pod="openshift-marketplace/redhat-marketplace-q6m25" Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.129234 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a4df431-deee-45d6-9c07-7a1616a23d91-utilities\") pod \"redhat-marketplace-q6m25\" (UID: \"0a4df431-deee-45d6-9c07-7a1616a23d91\") " pod="openshift-marketplace/redhat-marketplace-q6m25" Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.129288 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a4df431-deee-45d6-9c07-7a1616a23d91-catalog-content\") pod \"redhat-marketplace-q6m25\" (UID: \"0a4df431-deee-45d6-9c07-7a1616a23d91\") " pod="openshift-marketplace/redhat-marketplace-q6m25" Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.129318 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6efe6a93-4ea8-4383-b104-988fc603d3ef-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"6efe6a93-4ea8-4383-b104-988fc603d3ef\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.150021 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8p5ww\" (UniqueName: \"kubernetes.io/projected/0a4df431-deee-45d6-9c07-7a1616a23d91-kube-api-access-8p5ww\") pod \"redhat-marketplace-q6m25\" (UID: \"0a4df431-deee-45d6-9c07-7a1616a23d91\") " pod="openshift-marketplace/redhat-marketplace-q6m25" Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.163474 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6efe6a93-4ea8-4383-b104-988fc603d3ef-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"6efe6a93-4ea8-4383-b104-988fc603d3ef\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.175145 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.200578 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 29 04:40:23 crc kubenswrapper[4799]: W1129 04:40:23.234657 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod2a4dbe70_2ff5_41cb_8d0a_02e4451fca96.slice/crio-0d97e26f1ba09bc2e541849ec3b25c2e04d6c079b033df06d31ce93e7579a800 WatchSource:0}: Error finding container 0d97e26f1ba09bc2e541849ec3b25c2e04d6c079b033df06d31ce93e7579a800: Status 404 returned error can't find the container with id 0d97e26f1ba09bc2e541849ec3b25c2e04d6c079b033df06d31ce93e7579a800 Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.270072 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q6m25" Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.305031 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" event={"ID":"48c7bbb5-36e2-4677-b259-d3da6302d415","Type":"ContainerStarted","Data":"945cf67f84af377435db6cb0366c38af62992b22dc85a2294b4e1536f1855ae7"} Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.306153 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.315418 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-crfwf"] Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.316519 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-crfwf" Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.330697 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-crfwf"] Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.337330 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"2a4dbe70-2ff5-41cb-8d0a-02e4451fca96","Type":"ContainerStarted","Data":"0d97e26f1ba09bc2e541849ec3b25c2e04d6c079b033df06d31ce93e7579a800"} Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.344653 4799 patch_prober.go:28] interesting pod/router-default-5444994796-jrhwp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 29 04:40:23 crc kubenswrapper[4799]: [-]has-synced failed: reason withheld Nov 29 04:40:23 crc kubenswrapper[4799]: [+]process-running ok Nov 29 04:40:23 crc kubenswrapper[4799]: healthz check failed Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.344700 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jrhwp" podUID="36a3c13c-2670-4b33-a9fe-9071b35365fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.358138 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" podStartSLOduration=74.358115179 podStartE2EDuration="1m14.358115179s" podCreationTimestamp="2025-11-29 04:39:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:23.334527201 +0000 UTC m=+98.977457621" watchObservedRunningTime="2025-11-29 04:40:23.358115179 +0000 UTC m=+99.001045579" Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.436613 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0efffb4f-375f-49b8-8ff8-cc4149f71847-catalog-content\") pod \"redhat-marketplace-crfwf\" (UID: \"0efffb4f-375f-49b8-8ff8-cc4149f71847\") " pod="openshift-marketplace/redhat-marketplace-crfwf" Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.436702 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzsdx\" (UniqueName: \"kubernetes.io/projected/0efffb4f-375f-49b8-8ff8-cc4149f71847-kube-api-access-pzsdx\") pod \"redhat-marketplace-crfwf\" (UID: \"0efffb4f-375f-49b8-8ff8-cc4149f71847\") " pod="openshift-marketplace/redhat-marketplace-crfwf" Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.436757 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0efffb4f-375f-49b8-8ff8-cc4149f71847-utilities\") pod \"redhat-marketplace-crfwf\" (UID: \"0efffb4f-375f-49b8-8ff8-cc4149f71847\") " pod="openshift-marketplace/redhat-marketplace-crfwf" Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.539128 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzsdx\" (UniqueName: \"kubernetes.io/projected/0efffb4f-375f-49b8-8ff8-cc4149f71847-kube-api-access-pzsdx\") pod \"redhat-marketplace-crfwf\" (UID: \"0efffb4f-375f-49b8-8ff8-cc4149f71847\") " pod="openshift-marketplace/redhat-marketplace-crfwf" Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.539191 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0efffb4f-375f-49b8-8ff8-cc4149f71847-utilities\") pod \"redhat-marketplace-crfwf\" (UID: \"0efffb4f-375f-49b8-8ff8-cc4149f71847\") " pod="openshift-marketplace/redhat-marketplace-crfwf" Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.539254 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0efffb4f-375f-49b8-8ff8-cc4149f71847-catalog-content\") pod \"redhat-marketplace-crfwf\" (UID: \"0efffb4f-375f-49b8-8ff8-cc4149f71847\") " pod="openshift-marketplace/redhat-marketplace-crfwf" Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.540241 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0efffb4f-375f-49b8-8ff8-cc4149f71847-utilities\") pod \"redhat-marketplace-crfwf\" (UID: \"0efffb4f-375f-49b8-8ff8-cc4149f71847\") " pod="openshift-marketplace/redhat-marketplace-crfwf" Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.540345 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0efffb4f-375f-49b8-8ff8-cc4149f71847-catalog-content\") pod \"redhat-marketplace-crfwf\" (UID: \"0efffb4f-375f-49b8-8ff8-cc4149f71847\") " pod="openshift-marketplace/redhat-marketplace-crfwf" Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.559811 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzsdx\" (UniqueName: \"kubernetes.io/projected/0efffb4f-375f-49b8-8ff8-cc4149f71847-kube-api-access-pzsdx\") pod \"redhat-marketplace-crfwf\" (UID: \"0efffb4f-375f-49b8-8ff8-cc4149f71847\") " pod="openshift-marketplace/redhat-marketplace-crfwf" Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.643055 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-crfwf" Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.664758 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406510-f2lz7" Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.700075 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 29 04:40:23 crc kubenswrapper[4799]: W1129 04:40:23.702993 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod6efe6a93_4ea8_4383_b104_988fc603d3ef.slice/crio-a0b4c3f65fa3256aa46cef85f138c4b57fc4f170059392eeddac2a6ea01e3bf9 WatchSource:0}: Error finding container a0b4c3f65fa3256aa46cef85f138c4b57fc4f170059392eeddac2a6ea01e3bf9: Status 404 returned error can't find the container with id a0b4c3f65fa3256aa46cef85f138c4b57fc4f170059392eeddac2a6ea01e3bf9 Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.732329 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q6m25"] Nov 29 04:40:23 crc kubenswrapper[4799]: W1129 04:40:23.775312 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0a4df431_deee_45d6_9c07_7a1616a23d91.slice/crio-18dca1cad7a736e26a87a4a2d9e1e92da1cff0cc3e1e220c2b42e8e26b2058ce WatchSource:0}: Error finding container 18dca1cad7a736e26a87a4a2d9e1e92da1cff0cc3e1e220c2b42e8e26b2058ce: Status 404 returned error can't find the container with id 18dca1cad7a736e26a87a4a2d9e1e92da1cff0cc3e1e220c2b42e8e26b2058ce Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.850264 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/96a1e39b-0018-41b8-8d7c-9fa7697858e9-config-volume\") pod \"96a1e39b-0018-41b8-8d7c-9fa7697858e9\" (UID: \"96a1e39b-0018-41b8-8d7c-9fa7697858e9\") " Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.850433 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/96a1e39b-0018-41b8-8d7c-9fa7697858e9-secret-volume\") pod \"96a1e39b-0018-41b8-8d7c-9fa7697858e9\" (UID: \"96a1e39b-0018-41b8-8d7c-9fa7697858e9\") " Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.850509 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-svwvw\" (UniqueName: \"kubernetes.io/projected/96a1e39b-0018-41b8-8d7c-9fa7697858e9-kube-api-access-svwvw\") pod \"96a1e39b-0018-41b8-8d7c-9fa7697858e9\" (UID: \"96a1e39b-0018-41b8-8d7c-9fa7697858e9\") " Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.852197 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96a1e39b-0018-41b8-8d7c-9fa7697858e9-config-volume" (OuterVolumeSpecName: "config-volume") pod "96a1e39b-0018-41b8-8d7c-9fa7697858e9" (UID: "96a1e39b-0018-41b8-8d7c-9fa7697858e9"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.857233 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96a1e39b-0018-41b8-8d7c-9fa7697858e9-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "96a1e39b-0018-41b8-8d7c-9fa7697858e9" (UID: "96a1e39b-0018-41b8-8d7c-9fa7697858e9"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.857513 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96a1e39b-0018-41b8-8d7c-9fa7697858e9-kube-api-access-svwvw" (OuterVolumeSpecName: "kube-api-access-svwvw") pod "96a1e39b-0018-41b8-8d7c-9fa7697858e9" (UID: "96a1e39b-0018-41b8-8d7c-9fa7697858e9"). InnerVolumeSpecName "kube-api-access-svwvw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.918183 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-br2p6"] Nov 29 04:40:23 crc kubenswrapper[4799]: E1129 04:40:23.918646 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96a1e39b-0018-41b8-8d7c-9fa7697858e9" containerName="collect-profiles" Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.918668 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="96a1e39b-0018-41b8-8d7c-9fa7697858e9" containerName="collect-profiles" Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.919413 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="96a1e39b-0018-41b8-8d7c-9fa7697858e9" containerName="collect-profiles" Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.925999 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-br2p6"] Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.926178 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-br2p6" Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.928604 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.952285 4799 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/96a1e39b-0018-41b8-8d7c-9fa7697858e9-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.952314 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-svwvw\" (UniqueName: \"kubernetes.io/projected/96a1e39b-0018-41b8-8d7c-9fa7697858e9-kube-api-access-svwvw\") on node \"crc\" DevicePath \"\"" Nov 29 04:40:23 crc kubenswrapper[4799]: I1129 04:40:23.952324 4799 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/96a1e39b-0018-41b8-8d7c-9fa7697858e9-config-volume\") on node \"crc\" DevicePath \"\"" Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.020035 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-bbm46" Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.054023 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/390f5da3-eb7e-48af-bcc0-666c9172403e-utilities\") pod \"redhat-operators-br2p6\" (UID: \"390f5da3-eb7e-48af-bcc0-666c9172403e\") " pod="openshift-marketplace/redhat-operators-br2p6" Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.054121 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5b4sk\" (UniqueName: \"kubernetes.io/projected/390f5da3-eb7e-48af-bcc0-666c9172403e-kube-api-access-5b4sk\") pod \"redhat-operators-br2p6\" (UID: \"390f5da3-eb7e-48af-bcc0-666c9172403e\") " pod="openshift-marketplace/redhat-operators-br2p6" Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.054169 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/390f5da3-eb7e-48af-bcc0-666c9172403e-catalog-content\") pod \"redhat-operators-br2p6\" (UID: \"390f5da3-eb7e-48af-bcc0-666c9172403e\") " pod="openshift-marketplace/redhat-operators-br2p6" Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.060464 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-rcsd7" Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.060504 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-rcsd7" Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.072848 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-crfwf"] Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.078666 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-rcsd7" Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.081910 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zv6wk" Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.081938 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zv6wk" Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.088337 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zv6wk" Nov 29 04:40:24 crc kubenswrapper[4799]: W1129 04:40:24.116411 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0efffb4f_375f_49b8_8ff8_cc4149f71847.slice/crio-0ef24046dbe9b6f9e6c24d89f656148f1cb602813654ca6935f1b84d18824e24 WatchSource:0}: Error finding container 0ef24046dbe9b6f9e6c24d89f656148f1cb602813654ca6935f1b84d18824e24: Status 404 returned error can't find the container with id 0ef24046dbe9b6f9e6c24d89f656148f1cb602813654ca6935f1b84d18824e24 Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.155910 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5b4sk\" (UniqueName: \"kubernetes.io/projected/390f5da3-eb7e-48af-bcc0-666c9172403e-kube-api-access-5b4sk\") pod \"redhat-operators-br2p6\" (UID: \"390f5da3-eb7e-48af-bcc0-666c9172403e\") " pod="openshift-marketplace/redhat-operators-br2p6" Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.156741 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/390f5da3-eb7e-48af-bcc0-666c9172403e-catalog-content\") pod \"redhat-operators-br2p6\" (UID: \"390f5da3-eb7e-48af-bcc0-666c9172403e\") " pod="openshift-marketplace/redhat-operators-br2p6" Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.156878 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/390f5da3-eb7e-48af-bcc0-666c9172403e-utilities\") pod \"redhat-operators-br2p6\" (UID: \"390f5da3-eb7e-48af-bcc0-666c9172403e\") " pod="openshift-marketplace/redhat-operators-br2p6" Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.159574 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/390f5da3-eb7e-48af-bcc0-666c9172403e-utilities\") pod \"redhat-operators-br2p6\" (UID: \"390f5da3-eb7e-48af-bcc0-666c9172403e\") " pod="openshift-marketplace/redhat-operators-br2p6" Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.160364 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/390f5da3-eb7e-48af-bcc0-666c9172403e-catalog-content\") pod \"redhat-operators-br2p6\" (UID: \"390f5da3-eb7e-48af-bcc0-666c9172403e\") " pod="openshift-marketplace/redhat-operators-br2p6" Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.182707 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5b4sk\" (UniqueName: \"kubernetes.io/projected/390f5da3-eb7e-48af-bcc0-666c9172403e-kube-api-access-5b4sk\") pod \"redhat-operators-br2p6\" (UID: \"390f5da3-eb7e-48af-bcc0-666c9172403e\") " pod="openshift-marketplace/redhat-operators-br2p6" Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.255722 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-br2p6" Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.278283 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-2qb64" Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.278323 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-2qb64" Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.280747 4799 patch_prober.go:28] interesting pod/console-f9d7485db-2qb64 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.29:8443/health\": dial tcp 10.217.0.29:8443: connect: connection refused" start-of-body= Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.280816 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-2qb64" podUID="b6794cf7-9161-499d-ba18-74b95caaf3ae" containerName="console" probeResult="failure" output="Get \"https://10.217.0.29:8443/health\": dial tcp 10.217.0.29:8443: connect: connection refused" Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.312328 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xlxlr"] Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.315369 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xlxlr" Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.327810 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xlxlr"] Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.339913 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-jrhwp" Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.343231 4799 patch_prober.go:28] interesting pod/router-default-5444994796-jrhwp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 29 04:40:24 crc kubenswrapper[4799]: [-]has-synced failed: reason withheld Nov 29 04:40:24 crc kubenswrapper[4799]: [+]process-running ok Nov 29 04:40:24 crc kubenswrapper[4799]: healthz check failed Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.343300 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jrhwp" podUID="36a3c13c-2670-4b33-a9fe-9071b35365fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.355844 4799 generic.go:334] "Generic (PLEG): container finished" podID="0a4df431-deee-45d6-9c07-7a1616a23d91" containerID="1ad6d883f0343277fb0d10110b42ad0e79a341d9e6e433adf083700d7d30b40a" exitCode=0 Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.355901 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q6m25" event={"ID":"0a4df431-deee-45d6-9c07-7a1616a23d91","Type":"ContainerDied","Data":"1ad6d883f0343277fb0d10110b42ad0e79a341d9e6e433adf083700d7d30b40a"} Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.355924 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q6m25" event={"ID":"0a4df431-deee-45d6-9c07-7a1616a23d91","Type":"ContainerStarted","Data":"18dca1cad7a736e26a87a4a2d9e1e92da1cff0cc3e1e220c2b42e8e26b2058ce"} Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.358251 4799 generic.go:334] "Generic (PLEG): container finished" podID="0efffb4f-375f-49b8-8ff8-cc4149f71847" containerID="9657fd890f3b8b97e14129e9f4ec54425883bd680d8da9fdad6333ffc22764dc" exitCode=0 Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.358304 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-crfwf" event={"ID":"0efffb4f-375f-49b8-8ff8-cc4149f71847","Type":"ContainerDied","Data":"9657fd890f3b8b97e14129e9f4ec54425883bd680d8da9fdad6333ffc22764dc"} Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.358320 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-crfwf" event={"ID":"0efffb4f-375f-49b8-8ff8-cc4149f71847","Type":"ContainerStarted","Data":"0ef24046dbe9b6f9e6c24d89f656148f1cb602813654ca6935f1b84d18824e24"} Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.360542 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"2a4dbe70-2ff5-41cb-8d0a-02e4451fca96","Type":"ContainerStarted","Data":"2893ab366660f0234c2f65e067d2699dab8294495b868545b87065bf2ebde9e6"} Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.363019 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29406510-f2lz7" event={"ID":"96a1e39b-0018-41b8-8d7c-9fa7697858e9","Type":"ContainerDied","Data":"73a3ef287032122637ca8b076df9af0da790e3bd7b0ca554cf8f473d569aff12"} Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.363040 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="73a3ef287032122637ca8b076df9af0da790e3bd7b0ca554cf8f473d569aff12" Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.363088 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406510-f2lz7" Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.380323 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"6efe6a93-4ea8-4383-b104-988fc603d3ef","Type":"ContainerStarted","Data":"a0b4c3f65fa3256aa46cef85f138c4b57fc4f170059392eeddac2a6ea01e3bf9"} Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.388150 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zv6wk" Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.393658 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-rcsd7" Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.421567 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.421546294 podStartE2EDuration="2.421546294s" podCreationTimestamp="2025-11-29 04:40:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:24.418512996 +0000 UTC m=+100.061443396" watchObservedRunningTime="2025-11-29 04:40:24.421546294 +0000 UTC m=+100.064476694" Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.470559 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecc63394-2cf6-4a33-8e1b-a02042b8da4f-catalog-content\") pod \"redhat-operators-xlxlr\" (UID: \"ecc63394-2cf6-4a33-8e1b-a02042b8da4f\") " pod="openshift-marketplace/redhat-operators-xlxlr" Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.473603 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smhtx\" (UniqueName: \"kubernetes.io/projected/ecc63394-2cf6-4a33-8e1b-a02042b8da4f-kube-api-access-smhtx\") pod \"redhat-operators-xlxlr\" (UID: \"ecc63394-2cf6-4a33-8e1b-a02042b8da4f\") " pod="openshift-marketplace/redhat-operators-xlxlr" Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.473710 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecc63394-2cf6-4a33-8e1b-a02042b8da4f-utilities\") pod \"redhat-operators-xlxlr\" (UID: \"ecc63394-2cf6-4a33-8e1b-a02042b8da4f\") " pod="openshift-marketplace/redhat-operators-xlxlr" Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.575439 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smhtx\" (UniqueName: \"kubernetes.io/projected/ecc63394-2cf6-4a33-8e1b-a02042b8da4f-kube-api-access-smhtx\") pod \"redhat-operators-xlxlr\" (UID: \"ecc63394-2cf6-4a33-8e1b-a02042b8da4f\") " pod="openshift-marketplace/redhat-operators-xlxlr" Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.575583 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecc63394-2cf6-4a33-8e1b-a02042b8da4f-utilities\") pod \"redhat-operators-xlxlr\" (UID: \"ecc63394-2cf6-4a33-8e1b-a02042b8da4f\") " pod="openshift-marketplace/redhat-operators-xlxlr" Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.575660 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecc63394-2cf6-4a33-8e1b-a02042b8da4f-catalog-content\") pod \"redhat-operators-xlxlr\" (UID: \"ecc63394-2cf6-4a33-8e1b-a02042b8da4f\") " pod="openshift-marketplace/redhat-operators-xlxlr" Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.578140 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecc63394-2cf6-4a33-8e1b-a02042b8da4f-catalog-content\") pod \"redhat-operators-xlxlr\" (UID: \"ecc63394-2cf6-4a33-8e1b-a02042b8da4f\") " pod="openshift-marketplace/redhat-operators-xlxlr" Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.579154 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecc63394-2cf6-4a33-8e1b-a02042b8da4f-utilities\") pod \"redhat-operators-xlxlr\" (UID: \"ecc63394-2cf6-4a33-8e1b-a02042b8da4f\") " pod="openshift-marketplace/redhat-operators-xlxlr" Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.630199 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smhtx\" (UniqueName: \"kubernetes.io/projected/ecc63394-2cf6-4a33-8e1b-a02042b8da4f-kube-api-access-smhtx\") pod \"redhat-operators-xlxlr\" (UID: \"ecc63394-2cf6-4a33-8e1b-a02042b8da4f\") " pod="openshift-marketplace/redhat-operators-xlxlr" Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.680576 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-br2p6"] Nov 29 04:40:24 crc kubenswrapper[4799]: I1129 04:40:24.829996 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xlxlr" Nov 29 04:40:25 crc kubenswrapper[4799]: I1129 04:40:25.344460 4799 patch_prober.go:28] interesting pod/router-default-5444994796-jrhwp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 29 04:40:25 crc kubenswrapper[4799]: [-]has-synced failed: reason withheld Nov 29 04:40:25 crc kubenswrapper[4799]: [+]process-running ok Nov 29 04:40:25 crc kubenswrapper[4799]: healthz check failed Nov 29 04:40:25 crc kubenswrapper[4799]: I1129 04:40:25.344980 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jrhwp" podUID="36a3c13c-2670-4b33-a9fe-9071b35365fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 29 04:40:25 crc kubenswrapper[4799]: I1129 04:40:25.387341 4799 generic.go:334] "Generic (PLEG): container finished" podID="390f5da3-eb7e-48af-bcc0-666c9172403e" containerID="dcf896bbfd487aee632ced936832d741251434c5f231ea563fbf32d3084cad97" exitCode=0 Nov 29 04:40:25 crc kubenswrapper[4799]: I1129 04:40:25.387428 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-br2p6" event={"ID":"390f5da3-eb7e-48af-bcc0-666c9172403e","Type":"ContainerDied","Data":"dcf896bbfd487aee632ced936832d741251434c5f231ea563fbf32d3084cad97"} Nov 29 04:40:25 crc kubenswrapper[4799]: I1129 04:40:25.387458 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-br2p6" event={"ID":"390f5da3-eb7e-48af-bcc0-666c9172403e","Type":"ContainerStarted","Data":"07a5bc7f0abf4a702dd4d34fd6090cc33b820ce328d7c0a2854b6146fe98a8ab"} Nov 29 04:40:25 crc kubenswrapper[4799]: I1129 04:40:25.390974 4799 generic.go:334] "Generic (PLEG): container finished" podID="6efe6a93-4ea8-4383-b104-988fc603d3ef" containerID="28c884ae739e9a8b6dddb36ca6d09498ca5cbd1ab379e216bc3cc717bc6112a9" exitCode=0 Nov 29 04:40:25 crc kubenswrapper[4799]: I1129 04:40:25.391064 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"6efe6a93-4ea8-4383-b104-988fc603d3ef","Type":"ContainerDied","Data":"28c884ae739e9a8b6dddb36ca6d09498ca5cbd1ab379e216bc3cc717bc6112a9"} Nov 29 04:40:25 crc kubenswrapper[4799]: I1129 04:40:25.396688 4799 generic.go:334] "Generic (PLEG): container finished" podID="2a4dbe70-2ff5-41cb-8d0a-02e4451fca96" containerID="2893ab366660f0234c2f65e067d2699dab8294495b868545b87065bf2ebde9e6" exitCode=0 Nov 29 04:40:25 crc kubenswrapper[4799]: I1129 04:40:25.396847 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"2a4dbe70-2ff5-41cb-8d0a-02e4451fca96","Type":"ContainerDied","Data":"2893ab366660f0234c2f65e067d2699dab8294495b868545b87065bf2ebde9e6"} Nov 29 04:40:25 crc kubenswrapper[4799]: I1129 04:40:25.512103 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xlxlr"] Nov 29 04:40:25 crc kubenswrapper[4799]: I1129 04:40:25.684191 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Nov 29 04:40:26 crc kubenswrapper[4799]: I1129 04:40:26.345353 4799 patch_prober.go:28] interesting pod/router-default-5444994796-jrhwp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 29 04:40:26 crc kubenswrapper[4799]: [-]has-synced failed: reason withheld Nov 29 04:40:26 crc kubenswrapper[4799]: [+]process-running ok Nov 29 04:40:26 crc kubenswrapper[4799]: healthz check failed Nov 29 04:40:26 crc kubenswrapper[4799]: I1129 04:40:26.345889 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jrhwp" podUID="36a3c13c-2670-4b33-a9fe-9071b35365fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 29 04:40:26 crc kubenswrapper[4799]: I1129 04:40:26.492023 4799 generic.go:334] "Generic (PLEG): container finished" podID="ecc63394-2cf6-4a33-8e1b-a02042b8da4f" containerID="63c5e2dbdef44c5a12eef3bd8d44a34373e246e56f182fdd9c84c1823c5c0dc3" exitCode=0 Nov 29 04:40:26 crc kubenswrapper[4799]: I1129 04:40:26.492094 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xlxlr" event={"ID":"ecc63394-2cf6-4a33-8e1b-a02042b8da4f","Type":"ContainerDied","Data":"63c5e2dbdef44c5a12eef3bd8d44a34373e246e56f182fdd9c84c1823c5c0dc3"} Nov 29 04:40:26 crc kubenswrapper[4799]: I1129 04:40:26.492194 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xlxlr" event={"ID":"ecc63394-2cf6-4a33-8e1b-a02042b8da4f","Type":"ContainerStarted","Data":"0aaa9681602c482a11c0da6b44cdb470ea19003b6981838a973b0bf11877fd1a"} Nov 29 04:40:26 crc kubenswrapper[4799]: I1129 04:40:26.968783 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 29 04:40:26 crc kubenswrapper[4799]: I1129 04:40:26.983339 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=1.983322254 podStartE2EDuration="1.983322254s" podCreationTimestamp="2025-11-29 04:40:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:26.554362666 +0000 UTC m=+102.197293066" watchObservedRunningTime="2025-11-29 04:40:26.983322254 +0000 UTC m=+102.626252654" Nov 29 04:40:27 crc kubenswrapper[4799]: I1129 04:40:27.035398 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6efe6a93-4ea8-4383-b104-988fc603d3ef-kubelet-dir\") pod \"6efe6a93-4ea8-4383-b104-988fc603d3ef\" (UID: \"6efe6a93-4ea8-4383-b104-988fc603d3ef\") " Nov 29 04:40:27 crc kubenswrapper[4799]: I1129 04:40:27.035522 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6efe6a93-4ea8-4383-b104-988fc603d3ef-kube-api-access\") pod \"6efe6a93-4ea8-4383-b104-988fc603d3ef\" (UID: \"6efe6a93-4ea8-4383-b104-988fc603d3ef\") " Nov 29 04:40:27 crc kubenswrapper[4799]: I1129 04:40:27.035687 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6efe6a93-4ea8-4383-b104-988fc603d3ef-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "6efe6a93-4ea8-4383-b104-988fc603d3ef" (UID: "6efe6a93-4ea8-4383-b104-988fc603d3ef"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 04:40:27 crc kubenswrapper[4799]: I1129 04:40:27.035961 4799 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6efe6a93-4ea8-4383-b104-988fc603d3ef-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 29 04:40:27 crc kubenswrapper[4799]: I1129 04:40:27.044697 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6efe6a93-4ea8-4383-b104-988fc603d3ef-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "6efe6a93-4ea8-4383-b104-988fc603d3ef" (UID: "6efe6a93-4ea8-4383-b104-988fc603d3ef"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:40:27 crc kubenswrapper[4799]: I1129 04:40:27.068941 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 29 04:40:27 crc kubenswrapper[4799]: I1129 04:40:27.136504 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2a4dbe70-2ff5-41cb-8d0a-02e4451fca96-kube-api-access\") pod \"2a4dbe70-2ff5-41cb-8d0a-02e4451fca96\" (UID: \"2a4dbe70-2ff5-41cb-8d0a-02e4451fca96\") " Nov 29 04:40:27 crc kubenswrapper[4799]: I1129 04:40:27.136612 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2a4dbe70-2ff5-41cb-8d0a-02e4451fca96-kubelet-dir\") pod \"2a4dbe70-2ff5-41cb-8d0a-02e4451fca96\" (UID: \"2a4dbe70-2ff5-41cb-8d0a-02e4451fca96\") " Nov 29 04:40:27 crc kubenswrapper[4799]: I1129 04:40:27.137051 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6efe6a93-4ea8-4383-b104-988fc603d3ef-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 29 04:40:27 crc kubenswrapper[4799]: I1129 04:40:27.137104 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2a4dbe70-2ff5-41cb-8d0a-02e4451fca96-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "2a4dbe70-2ff5-41cb-8d0a-02e4451fca96" (UID: "2a4dbe70-2ff5-41cb-8d0a-02e4451fca96"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 04:40:27 crc kubenswrapper[4799]: I1129 04:40:27.141099 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a4dbe70-2ff5-41cb-8d0a-02e4451fca96-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "2a4dbe70-2ff5-41cb-8d0a-02e4451fca96" (UID: "2a4dbe70-2ff5-41cb-8d0a-02e4451fca96"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:40:27 crc kubenswrapper[4799]: I1129 04:40:27.238409 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2a4dbe70-2ff5-41cb-8d0a-02e4451fca96-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 29 04:40:27 crc kubenswrapper[4799]: I1129 04:40:27.238435 4799 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2a4dbe70-2ff5-41cb-8d0a-02e4451fca96-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 29 04:40:27 crc kubenswrapper[4799]: I1129 04:40:27.345832 4799 patch_prober.go:28] interesting pod/router-default-5444994796-jrhwp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 29 04:40:27 crc kubenswrapper[4799]: [-]has-synced failed: reason withheld Nov 29 04:40:27 crc kubenswrapper[4799]: [+]process-running ok Nov 29 04:40:27 crc kubenswrapper[4799]: healthz check failed Nov 29 04:40:27 crc kubenswrapper[4799]: I1129 04:40:27.346514 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jrhwp" podUID="36a3c13c-2670-4b33-a9fe-9071b35365fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 29 04:40:27 crc kubenswrapper[4799]: I1129 04:40:27.506239 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 29 04:40:27 crc kubenswrapper[4799]: I1129 04:40:27.506328 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"2a4dbe70-2ff5-41cb-8d0a-02e4451fca96","Type":"ContainerDied","Data":"0d97e26f1ba09bc2e541849ec3b25c2e04d6c079b033df06d31ce93e7579a800"} Nov 29 04:40:27 crc kubenswrapper[4799]: I1129 04:40:27.506415 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d97e26f1ba09bc2e541849ec3b25c2e04d6c079b033df06d31ce93e7579a800" Nov 29 04:40:27 crc kubenswrapper[4799]: I1129 04:40:27.510330 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 29 04:40:27 crc kubenswrapper[4799]: I1129 04:40:27.510316 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"6efe6a93-4ea8-4383-b104-988fc603d3ef","Type":"ContainerDied","Data":"a0b4c3f65fa3256aa46cef85f138c4b57fc4f170059392eeddac2a6ea01e3bf9"} Nov 29 04:40:27 crc kubenswrapper[4799]: I1129 04:40:27.510454 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a0b4c3f65fa3256aa46cef85f138c4b57fc4f170059392eeddac2a6ea01e3bf9" Nov 29 04:40:27 crc kubenswrapper[4799]: I1129 04:40:27.947748 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ce8ce318-27e6-434d-96c8-591d250262fd-metrics-certs\") pod \"network-metrics-daemon-qx9wn\" (UID: \"ce8ce318-27e6-434d-96c8-591d250262fd\") " pod="openshift-multus/network-metrics-daemon-qx9wn" Nov 29 04:40:27 crc kubenswrapper[4799]: I1129 04:40:27.970479 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ce8ce318-27e6-434d-96c8-591d250262fd-metrics-certs\") pod \"network-metrics-daemon-qx9wn\" (UID: \"ce8ce318-27e6-434d-96c8-591d250262fd\") " pod="openshift-multus/network-metrics-daemon-qx9wn" Nov 29 04:40:28 crc kubenswrapper[4799]: I1129 04:40:28.090010 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qx9wn" Nov 29 04:40:28 crc kubenswrapper[4799]: I1129 04:40:28.345229 4799 patch_prober.go:28] interesting pod/router-default-5444994796-jrhwp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 29 04:40:28 crc kubenswrapper[4799]: [-]has-synced failed: reason withheld Nov 29 04:40:28 crc kubenswrapper[4799]: [+]process-running ok Nov 29 04:40:28 crc kubenswrapper[4799]: healthz check failed Nov 29 04:40:28 crc kubenswrapper[4799]: I1129 04:40:28.345765 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jrhwp" podUID="36a3c13c-2670-4b33-a9fe-9071b35365fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 29 04:40:28 crc kubenswrapper[4799]: I1129 04:40:28.591519 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-qx9wn"] Nov 29 04:40:28 crc kubenswrapper[4799]: W1129 04:40:28.632625 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podce8ce318_27e6_434d_96c8_591d250262fd.slice/crio-334013ea3fedb198931b3617fa3389e253c2d566b87618f00a377eba2279665c WatchSource:0}: Error finding container 334013ea3fedb198931b3617fa3389e253c2d566b87618f00a377eba2279665c: Status 404 returned error can't find the container with id 334013ea3fedb198931b3617fa3389e253c2d566b87618f00a377eba2279665c Nov 29 04:40:29 crc kubenswrapper[4799]: I1129 04:40:29.343394 4799 patch_prober.go:28] interesting pod/router-default-5444994796-jrhwp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 29 04:40:29 crc kubenswrapper[4799]: [-]has-synced failed: reason withheld Nov 29 04:40:29 crc kubenswrapper[4799]: [+]process-running ok Nov 29 04:40:29 crc kubenswrapper[4799]: healthz check failed Nov 29 04:40:29 crc kubenswrapper[4799]: I1129 04:40:29.343527 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jrhwp" podUID="36a3c13c-2670-4b33-a9fe-9071b35365fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 29 04:40:29 crc kubenswrapper[4799]: I1129 04:40:29.527192 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-qx9wn" event={"ID":"ce8ce318-27e6-434d-96c8-591d250262fd","Type":"ContainerStarted","Data":"334013ea3fedb198931b3617fa3389e253c2d566b87618f00a377eba2279665c"} Nov 29 04:40:30 crc kubenswrapper[4799]: I1129 04:40:30.343990 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-jrhwp" Nov 29 04:40:30 crc kubenswrapper[4799]: I1129 04:40:30.347388 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-jrhwp" Nov 29 04:40:30 crc kubenswrapper[4799]: I1129 04:40:30.449877 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-vbffw" Nov 29 04:40:30 crc kubenswrapper[4799]: I1129 04:40:30.546938 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-qx9wn" event={"ID":"ce8ce318-27e6-434d-96c8-591d250262fd","Type":"ContainerStarted","Data":"1320d1fc03682c14f189556a8d94aef775810d2e5dab6f7f0c5c838dc7ee6010"} Nov 29 04:40:31 crc kubenswrapper[4799]: I1129 04:40:31.558131 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-qx9wn" event={"ID":"ce8ce318-27e6-434d-96c8-591d250262fd","Type":"ContainerStarted","Data":"c853c126d7277b67eb1a5f2f78500bcbff55cd1b0e425dcd3896c88b543426ef"} Nov 29 04:40:32 crc kubenswrapper[4799]: I1129 04:40:32.585648 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-qx9wn" podStartSLOduration=83.585624758 podStartE2EDuration="1m23.585624758s" podCreationTimestamp="2025-11-29 04:39:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:40:32.582433135 +0000 UTC m=+108.225363535" watchObservedRunningTime="2025-11-29 04:40:32.585624758 +0000 UTC m=+108.228555168" Nov 29 04:40:34 crc kubenswrapper[4799]: I1129 04:40:34.293381 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-2qb64" Nov 29 04:40:34 crc kubenswrapper[4799]: I1129 04:40:34.298241 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-2qb64" Nov 29 04:40:40 crc kubenswrapper[4799]: I1129 04:40:40.451741 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:40:41 crc kubenswrapper[4799]: I1129 04:40:41.924759 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:40:51 crc kubenswrapper[4799]: E1129 04:40:51.762581 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 29 04:40:51 crc kubenswrapper[4799]: E1129 04:40:51.763431 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-b78n8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-dnz56_openshift-marketplace(8734a259-eaa8-459e-b38d-e821e336a863): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 29 04:40:51 crc kubenswrapper[4799]: E1129 04:40:51.764657 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-dnz56" podUID="8734a259-eaa8-459e-b38d-e821e336a863" Nov 29 04:40:54 crc kubenswrapper[4799]: E1129 04:40:54.777929 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-dnz56" podUID="8734a259-eaa8-459e-b38d-e821e336a863" Nov 29 04:40:54 crc kubenswrapper[4799]: I1129 04:40:54.804391 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qd7d2" Nov 29 04:40:54 crc kubenswrapper[4799]: E1129 04:40:54.969122 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 29 04:40:54 crc kubenswrapper[4799]: E1129 04:40:54.969451 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5b4sk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-br2p6_openshift-marketplace(390f5da3-eb7e-48af-bcc0-666c9172403e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 29 04:40:54 crc kubenswrapper[4799]: E1129 04:40:54.970646 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-br2p6" podUID="390f5da3-eb7e-48af-bcc0-666c9172403e" Nov 29 04:40:55 crc kubenswrapper[4799]: E1129 04:40:55.890499 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 29 04:40:55 crc kubenswrapper[4799]: E1129 04:40:55.890722 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-d676x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-vl566_openshift-marketplace(b0bccaa4-e99f-47b1-bf20-2895f289ffd7): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 29 04:40:55 crc kubenswrapper[4799]: E1129 04:40:55.892948 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-vl566" podUID="b0bccaa4-e99f-47b1-bf20-2895f289ffd7" Nov 29 04:40:56 crc kubenswrapper[4799]: E1129 04:40:56.133436 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-br2p6" podUID="390f5da3-eb7e-48af-bcc0-666c9172403e" Nov 29 04:40:57 crc kubenswrapper[4799]: E1129 04:40:57.363068 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 29 04:40:57 crc kubenswrapper[4799]: E1129 04:40:57.363655 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pzsdx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-crfwf_openshift-marketplace(0efffb4f-375f-49b8-8ff8-cc4149f71847): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 29 04:40:57 crc kubenswrapper[4799]: E1129 04:40:57.365999 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-crfwf" podUID="0efffb4f-375f-49b8-8ff8-cc4149f71847" Nov 29 04:40:59 crc kubenswrapper[4799]: E1129 04:40:59.183228 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-crfwf" podUID="0efffb4f-375f-49b8-8ff8-cc4149f71847" Nov 29 04:40:59 crc kubenswrapper[4799]: E1129 04:40:59.184227 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-vl566" podUID="b0bccaa4-e99f-47b1-bf20-2895f289ffd7" Nov 29 04:40:59 crc kubenswrapper[4799]: E1129 04:40:59.248126 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 29 04:40:59 crc kubenswrapper[4799]: E1129 04:40:59.248387 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7xvpw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-mplj2_openshift-marketplace(8a2cc255-bc5f-4253-bc2b-e74817971afe): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 29 04:40:59 crc kubenswrapper[4799]: E1129 04:40:59.249576 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-mplj2" podUID="8a2cc255-bc5f-4253-bc2b-e74817971afe" Nov 29 04:40:59 crc kubenswrapper[4799]: E1129 04:40:59.317802 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 29 04:40:59 crc kubenswrapper[4799]: E1129 04:40:59.318042 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8p5ww,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-q6m25_openshift-marketplace(0a4df431-deee-45d6-9c07-7a1616a23d91): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 29 04:40:59 crc kubenswrapper[4799]: E1129 04:40:59.319303 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-q6m25" podUID="0a4df431-deee-45d6-9c07-7a1616a23d91" Nov 29 04:40:59 crc kubenswrapper[4799]: E1129 04:40:59.337013 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 29 04:40:59 crc kubenswrapper[4799]: E1129 04:40:59.337186 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-n6ssg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-jlvpn_openshift-marketplace(2544bf9f-2bb8-4661-997c-9b25a96103d4): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 29 04:40:59 crc kubenswrapper[4799]: E1129 04:40:59.338411 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-jlvpn" podUID="2544bf9f-2bb8-4661-997c-9b25a96103d4" Nov 29 04:40:59 crc kubenswrapper[4799]: I1129 04:40:59.765365 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xlxlr" event={"ID":"ecc63394-2cf6-4a33-8e1b-a02042b8da4f","Type":"ContainerStarted","Data":"0bd9fb54a0c5c77b88fc6c0a563c36893388c12864861531d5ec4ae408caef94"} Nov 29 04:40:59 crc kubenswrapper[4799]: E1129 04:40:59.767738 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-mplj2" podUID="8a2cc255-bc5f-4253-bc2b-e74817971afe" Nov 29 04:40:59 crc kubenswrapper[4799]: E1129 04:40:59.768145 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-jlvpn" podUID="2544bf9f-2bb8-4661-997c-9b25a96103d4" Nov 29 04:40:59 crc kubenswrapper[4799]: E1129 04:40:59.768644 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-q6m25" podUID="0a4df431-deee-45d6-9c07-7a1616a23d91" Nov 29 04:41:00 crc kubenswrapper[4799]: I1129 04:41:00.774403 4799 generic.go:334] "Generic (PLEG): container finished" podID="ecc63394-2cf6-4a33-8e1b-a02042b8da4f" containerID="0bd9fb54a0c5c77b88fc6c0a563c36893388c12864861531d5ec4ae408caef94" exitCode=0 Nov 29 04:41:00 crc kubenswrapper[4799]: I1129 04:41:00.774470 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xlxlr" event={"ID":"ecc63394-2cf6-4a33-8e1b-a02042b8da4f","Type":"ContainerDied","Data":"0bd9fb54a0c5c77b88fc6c0a563c36893388c12864861531d5ec4ae408caef94"} Nov 29 04:41:01 crc kubenswrapper[4799]: I1129 04:41:01.784074 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xlxlr" event={"ID":"ecc63394-2cf6-4a33-8e1b-a02042b8da4f","Type":"ContainerStarted","Data":"3ada1d3a01375e70ad1f7a656e7a48bc41707b0d852cddb11f2ce31cdfc6cb31"} Nov 29 04:41:01 crc kubenswrapper[4799]: I1129 04:41:01.811009 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xlxlr" podStartSLOduration=3.089637921 podStartE2EDuration="37.810973066s" podCreationTimestamp="2025-11-29 04:40:24 +0000 UTC" firstStartedPulling="2025-11-29 04:40:26.495307654 +0000 UTC m=+102.138238054" lastFinishedPulling="2025-11-29 04:41:01.216642759 +0000 UTC m=+136.859573199" observedRunningTime="2025-11-29 04:41:01.807342044 +0000 UTC m=+137.450272484" watchObservedRunningTime="2025-11-29 04:41:01.810973066 +0000 UTC m=+137.453903506" Nov 29 04:41:02 crc kubenswrapper[4799]: I1129 04:41:02.433903 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 29 04:41:02 crc kubenswrapper[4799]: E1129 04:41:02.434302 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6efe6a93-4ea8-4383-b104-988fc603d3ef" containerName="pruner" Nov 29 04:41:02 crc kubenswrapper[4799]: I1129 04:41:02.434320 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="6efe6a93-4ea8-4383-b104-988fc603d3ef" containerName="pruner" Nov 29 04:41:02 crc kubenswrapper[4799]: E1129 04:41:02.434348 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a4dbe70-2ff5-41cb-8d0a-02e4451fca96" containerName="pruner" Nov 29 04:41:02 crc kubenswrapper[4799]: I1129 04:41:02.434358 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a4dbe70-2ff5-41cb-8d0a-02e4451fca96" containerName="pruner" Nov 29 04:41:02 crc kubenswrapper[4799]: I1129 04:41:02.434492 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a4dbe70-2ff5-41cb-8d0a-02e4451fca96" containerName="pruner" Nov 29 04:41:02 crc kubenswrapper[4799]: I1129 04:41:02.434515 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="6efe6a93-4ea8-4383-b104-988fc603d3ef" containerName="pruner" Nov 29 04:41:02 crc kubenswrapper[4799]: I1129 04:41:02.435068 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 29 04:41:02 crc kubenswrapper[4799]: I1129 04:41:02.438916 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 29 04:41:02 crc kubenswrapper[4799]: I1129 04:41:02.439191 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 29 04:41:02 crc kubenswrapper[4799]: I1129 04:41:02.449291 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 29 04:41:02 crc kubenswrapper[4799]: I1129 04:41:02.602962 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2a096859-d13a-4791-b2d0-d59d452bc1d2-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"2a096859-d13a-4791-b2d0-d59d452bc1d2\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 29 04:41:02 crc kubenswrapper[4799]: I1129 04:41:02.603271 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2a096859-d13a-4791-b2d0-d59d452bc1d2-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"2a096859-d13a-4791-b2d0-d59d452bc1d2\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 29 04:41:02 crc kubenswrapper[4799]: I1129 04:41:02.704303 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2a096859-d13a-4791-b2d0-d59d452bc1d2-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"2a096859-d13a-4791-b2d0-d59d452bc1d2\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 29 04:41:02 crc kubenswrapper[4799]: I1129 04:41:02.704669 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2a096859-d13a-4791-b2d0-d59d452bc1d2-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"2a096859-d13a-4791-b2d0-d59d452bc1d2\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 29 04:41:02 crc kubenswrapper[4799]: I1129 04:41:02.704852 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2a096859-d13a-4791-b2d0-d59d452bc1d2-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"2a096859-d13a-4791-b2d0-d59d452bc1d2\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 29 04:41:02 crc kubenswrapper[4799]: I1129 04:41:02.734775 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2a096859-d13a-4791-b2d0-d59d452bc1d2-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"2a096859-d13a-4791-b2d0-d59d452bc1d2\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 29 04:41:02 crc kubenswrapper[4799]: I1129 04:41:02.753053 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 29 04:41:02 crc kubenswrapper[4799]: I1129 04:41:02.993338 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 29 04:41:03 crc kubenswrapper[4799]: I1129 04:41:03.802967 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"2a096859-d13a-4791-b2d0-d59d452bc1d2","Type":"ContainerStarted","Data":"b06d79331205a7435f4e7d1431d41be82faef062e01f652e9eed39f147b8f0ed"} Nov 29 04:41:03 crc kubenswrapper[4799]: I1129 04:41:03.803374 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"2a096859-d13a-4791-b2d0-d59d452bc1d2","Type":"ContainerStarted","Data":"e4058a517cca3b709fd1223d6cffc244f304f0c1e88e068255fd50df832ff892"} Nov 29 04:41:03 crc kubenswrapper[4799]: I1129 04:41:03.827282 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=1.8272500090000001 podStartE2EDuration="1.827250009s" podCreationTimestamp="2025-11-29 04:41:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:41:03.820948132 +0000 UTC m=+139.463878572" watchObservedRunningTime="2025-11-29 04:41:03.827250009 +0000 UTC m=+139.470180429" Nov 29 04:41:04 crc kubenswrapper[4799]: I1129 04:41:04.811937 4799 generic.go:334] "Generic (PLEG): container finished" podID="2a096859-d13a-4791-b2d0-d59d452bc1d2" containerID="b06d79331205a7435f4e7d1431d41be82faef062e01f652e9eed39f147b8f0ed" exitCode=0 Nov 29 04:41:04 crc kubenswrapper[4799]: I1129 04:41:04.812121 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"2a096859-d13a-4791-b2d0-d59d452bc1d2","Type":"ContainerDied","Data":"b06d79331205a7435f4e7d1431d41be82faef062e01f652e9eed39f147b8f0ed"} Nov 29 04:41:04 crc kubenswrapper[4799]: I1129 04:41:04.832141 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xlxlr" Nov 29 04:41:04 crc kubenswrapper[4799]: I1129 04:41:04.832225 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xlxlr" Nov 29 04:41:05 crc kubenswrapper[4799]: I1129 04:41:05.941157 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-xlxlr" podUID="ecc63394-2cf6-4a33-8e1b-a02042b8da4f" containerName="registry-server" probeResult="failure" output=< Nov 29 04:41:05 crc kubenswrapper[4799]: timeout: failed to connect service ":50051" within 1s Nov 29 04:41:05 crc kubenswrapper[4799]: > Nov 29 04:41:06 crc kubenswrapper[4799]: I1129 04:41:06.103004 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 29 04:41:06 crc kubenswrapper[4799]: I1129 04:41:06.176476 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2a096859-d13a-4791-b2d0-d59d452bc1d2-kubelet-dir\") pod \"2a096859-d13a-4791-b2d0-d59d452bc1d2\" (UID: \"2a096859-d13a-4791-b2d0-d59d452bc1d2\") " Nov 29 04:41:06 crc kubenswrapper[4799]: I1129 04:41:06.176561 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2a096859-d13a-4791-b2d0-d59d452bc1d2-kube-api-access\") pod \"2a096859-d13a-4791-b2d0-d59d452bc1d2\" (UID: \"2a096859-d13a-4791-b2d0-d59d452bc1d2\") " Nov 29 04:41:06 crc kubenswrapper[4799]: I1129 04:41:06.177622 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2a096859-d13a-4791-b2d0-d59d452bc1d2-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "2a096859-d13a-4791-b2d0-d59d452bc1d2" (UID: "2a096859-d13a-4791-b2d0-d59d452bc1d2"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 04:41:06 crc kubenswrapper[4799]: I1129 04:41:06.187892 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a096859-d13a-4791-b2d0-d59d452bc1d2-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "2a096859-d13a-4791-b2d0-d59d452bc1d2" (UID: "2a096859-d13a-4791-b2d0-d59d452bc1d2"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:41:06 crc kubenswrapper[4799]: I1129 04:41:06.277748 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2a096859-d13a-4791-b2d0-d59d452bc1d2-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 29 04:41:06 crc kubenswrapper[4799]: I1129 04:41:06.277781 4799 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2a096859-d13a-4791-b2d0-d59d452bc1d2-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 29 04:41:06 crc kubenswrapper[4799]: I1129 04:41:06.831016 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"2a096859-d13a-4791-b2d0-d59d452bc1d2","Type":"ContainerDied","Data":"e4058a517cca3b709fd1223d6cffc244f304f0c1e88e068255fd50df832ff892"} Nov 29 04:41:06 crc kubenswrapper[4799]: I1129 04:41:06.831473 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e4058a517cca3b709fd1223d6cffc244f304f0c1e88e068255fd50df832ff892" Nov 29 04:41:06 crc kubenswrapper[4799]: I1129 04:41:06.831271 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 29 04:41:09 crc kubenswrapper[4799]: I1129 04:41:09.833259 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 29 04:41:09 crc kubenswrapper[4799]: E1129 04:41:09.837938 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a096859-d13a-4791-b2d0-d59d452bc1d2" containerName="pruner" Nov 29 04:41:09 crc kubenswrapper[4799]: I1129 04:41:09.837992 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a096859-d13a-4791-b2d0-d59d452bc1d2" containerName="pruner" Nov 29 04:41:09 crc kubenswrapper[4799]: I1129 04:41:09.838271 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a096859-d13a-4791-b2d0-d59d452bc1d2" containerName="pruner" Nov 29 04:41:09 crc kubenswrapper[4799]: I1129 04:41:09.839272 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 29 04:41:09 crc kubenswrapper[4799]: I1129 04:41:09.843783 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 29 04:41:09 crc kubenswrapper[4799]: I1129 04:41:09.845327 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 29 04:41:09 crc kubenswrapper[4799]: I1129 04:41:09.845613 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 29 04:41:09 crc kubenswrapper[4799]: I1129 04:41:09.934360 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7dae4078-40b8-46a7-a83c-d162a1295c27-kube-api-access\") pod \"installer-9-crc\" (UID: \"7dae4078-40b8-46a7-a83c-d162a1295c27\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 29 04:41:09 crc kubenswrapper[4799]: I1129 04:41:09.934457 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/7dae4078-40b8-46a7-a83c-d162a1295c27-var-lock\") pod \"installer-9-crc\" (UID: \"7dae4078-40b8-46a7-a83c-d162a1295c27\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 29 04:41:09 crc kubenswrapper[4799]: I1129 04:41:09.934525 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7dae4078-40b8-46a7-a83c-d162a1295c27-kubelet-dir\") pod \"installer-9-crc\" (UID: \"7dae4078-40b8-46a7-a83c-d162a1295c27\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 29 04:41:10 crc kubenswrapper[4799]: I1129 04:41:10.035881 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7dae4078-40b8-46a7-a83c-d162a1295c27-kube-api-access\") pod \"installer-9-crc\" (UID: \"7dae4078-40b8-46a7-a83c-d162a1295c27\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 29 04:41:10 crc kubenswrapper[4799]: I1129 04:41:10.035977 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/7dae4078-40b8-46a7-a83c-d162a1295c27-var-lock\") pod \"installer-9-crc\" (UID: \"7dae4078-40b8-46a7-a83c-d162a1295c27\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 29 04:41:10 crc kubenswrapper[4799]: I1129 04:41:10.036039 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7dae4078-40b8-46a7-a83c-d162a1295c27-kubelet-dir\") pod \"installer-9-crc\" (UID: \"7dae4078-40b8-46a7-a83c-d162a1295c27\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 29 04:41:10 crc kubenswrapper[4799]: I1129 04:41:10.036157 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7dae4078-40b8-46a7-a83c-d162a1295c27-kubelet-dir\") pod \"installer-9-crc\" (UID: \"7dae4078-40b8-46a7-a83c-d162a1295c27\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 29 04:41:10 crc kubenswrapper[4799]: I1129 04:41:10.036180 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/7dae4078-40b8-46a7-a83c-d162a1295c27-var-lock\") pod \"installer-9-crc\" (UID: \"7dae4078-40b8-46a7-a83c-d162a1295c27\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 29 04:41:10 crc kubenswrapper[4799]: I1129 04:41:10.066168 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7dae4078-40b8-46a7-a83c-d162a1295c27-kube-api-access\") pod \"installer-9-crc\" (UID: \"7dae4078-40b8-46a7-a83c-d162a1295c27\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 29 04:41:10 crc kubenswrapper[4799]: I1129 04:41:10.158903 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 29 04:41:10 crc kubenswrapper[4799]: I1129 04:41:10.437305 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-vdxqp"] Nov 29 04:41:10 crc kubenswrapper[4799]: I1129 04:41:10.605589 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 29 04:41:10 crc kubenswrapper[4799]: W1129 04:41:10.614654 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod7dae4078_40b8_46a7_a83c_d162a1295c27.slice/crio-04d95f44ec2c81dd01c02f291e6a1d455016e35effdf1c9a6fb92558e72c4838 WatchSource:0}: Error finding container 04d95f44ec2c81dd01c02f291e6a1d455016e35effdf1c9a6fb92558e72c4838: Status 404 returned error can't find the container with id 04d95f44ec2c81dd01c02f291e6a1d455016e35effdf1c9a6fb92558e72c4838 Nov 29 04:41:10 crc kubenswrapper[4799]: I1129 04:41:10.858613 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"7dae4078-40b8-46a7-a83c-d162a1295c27","Type":"ContainerStarted","Data":"04d95f44ec2c81dd01c02f291e6a1d455016e35effdf1c9a6fb92558e72c4838"} Nov 29 04:41:11 crc kubenswrapper[4799]: I1129 04:41:11.769115 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 04:41:11 crc kubenswrapper[4799]: I1129 04:41:11.769861 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:41:11 crc kubenswrapper[4799]: I1129 04:41:11.770043 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 04:41:11 crc kubenswrapper[4799]: I1129 04:41:11.770265 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:41:11 crc kubenswrapper[4799]: I1129 04:41:11.771856 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 29 04:41:11 crc kubenswrapper[4799]: I1129 04:41:11.772388 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 29 04:41:11 crc kubenswrapper[4799]: I1129 04:41:11.772597 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 29 04:41:11 crc kubenswrapper[4799]: I1129 04:41:11.782885 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 29 04:41:11 crc kubenswrapper[4799]: I1129 04:41:11.790222 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:41:11 crc kubenswrapper[4799]: I1129 04:41:11.794020 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 04:41:11 crc kubenswrapper[4799]: I1129 04:41:11.794132 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 04:41:11 crc kubenswrapper[4799]: I1129 04:41:11.879009 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 04:41:11 crc kubenswrapper[4799]: I1129 04:41:11.896752 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 04:41:11 crc kubenswrapper[4799]: I1129 04:41:11.991058 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:41:12 crc kubenswrapper[4799]: I1129 04:41:12.173479 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 04:41:13 crc kubenswrapper[4799]: I1129 04:41:13.886175 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-br2p6" event={"ID":"390f5da3-eb7e-48af-bcc0-666c9172403e","Type":"ContainerStarted","Data":"ddde27ed7104ddebb76cfaca165251f8783fd7de7a5a7b91cd177cdc191cca55"} Nov 29 04:41:13 crc kubenswrapper[4799]: I1129 04:41:13.889848 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"2ef1b611425a9bc2dd0ee2f5f462b737562f882556d554294846b6c019e8a0d7"} Nov 29 04:41:13 crc kubenswrapper[4799]: I1129 04:41:13.893191 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dnz56" event={"ID":"8734a259-eaa8-459e-b38d-e821e336a863","Type":"ContainerStarted","Data":"728010b2c0c528c35fcfca45db3587f5630bf8ba9e9cad64ab24ca88cb02600e"} Nov 29 04:41:14 crc kubenswrapper[4799]: W1129 04:41:14.059223 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-e2554ae4d2fa5616ffaebe57209155b976dd6b95a075335121ac508759ed2840 WatchSource:0}: Error finding container e2554ae4d2fa5616ffaebe57209155b976dd6b95a075335121ac508759ed2840: Status 404 returned error can't find the container with id e2554ae4d2fa5616ffaebe57209155b976dd6b95a075335121ac508759ed2840 Nov 29 04:41:14 crc kubenswrapper[4799]: I1129 04:41:14.880646 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xlxlr" Nov 29 04:41:14 crc kubenswrapper[4799]: I1129 04:41:14.906144 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vl566" event={"ID":"b0bccaa4-e99f-47b1-bf20-2895f289ffd7","Type":"ContainerStarted","Data":"cf73277195389762deb95a92e3bc34711b22d193b6a676db03a9fc1a80c521b2"} Nov 29 04:41:14 crc kubenswrapper[4799]: I1129 04:41:14.912852 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jlvpn" event={"ID":"2544bf9f-2bb8-4661-997c-9b25a96103d4","Type":"ContainerStarted","Data":"9f6fff8b5ce57b269a0dd398202181c3a76dcc3aa7486e709184bf12d934965e"} Nov 29 04:41:14 crc kubenswrapper[4799]: I1129 04:41:14.925196 4799 generic.go:334] "Generic (PLEG): container finished" podID="390f5da3-eb7e-48af-bcc0-666c9172403e" containerID="ddde27ed7104ddebb76cfaca165251f8783fd7de7a5a7b91cd177cdc191cca55" exitCode=0 Nov 29 04:41:14 crc kubenswrapper[4799]: I1129 04:41:14.925271 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-br2p6" event={"ID":"390f5da3-eb7e-48af-bcc0-666c9172403e","Type":"ContainerDied","Data":"ddde27ed7104ddebb76cfaca165251f8783fd7de7a5a7b91cd177cdc191cca55"} Nov 29 04:41:14 crc kubenswrapper[4799]: I1129 04:41:14.934678 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xlxlr" Nov 29 04:41:14 crc kubenswrapper[4799]: I1129 04:41:14.939989 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"c2ed52ad2c4559b8171da76ed11f5d077eaa9d1295ff607ee7d4a10e9b12c124"} Nov 29 04:41:14 crc kubenswrapper[4799]: I1129 04:41:14.940050 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"17596db1f0cb9d56c0fd6fbae4c50af37386604e57ce492951d5c8ca45624a23"} Nov 29 04:41:14 crc kubenswrapper[4799]: I1129 04:41:14.944699 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"718365451ea77dee9ba09dae4bc5b11448f76eddc6e40d5a5b19c8ab48e4082d"} Nov 29 04:41:14 crc kubenswrapper[4799]: I1129 04:41:14.950606 4799 generic.go:334] "Generic (PLEG): container finished" podID="8734a259-eaa8-459e-b38d-e821e336a863" containerID="728010b2c0c528c35fcfca45db3587f5630bf8ba9e9cad64ab24ca88cb02600e" exitCode=0 Nov 29 04:41:14 crc kubenswrapper[4799]: I1129 04:41:14.950706 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dnz56" event={"ID":"8734a259-eaa8-459e-b38d-e821e336a863","Type":"ContainerDied","Data":"728010b2c0c528c35fcfca45db3587f5630bf8ba9e9cad64ab24ca88cb02600e"} Nov 29 04:41:14 crc kubenswrapper[4799]: I1129 04:41:14.957291 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"7dae4078-40b8-46a7-a83c-d162a1295c27","Type":"ContainerStarted","Data":"632a00da1e2b592990cf5d9f4cc02d6562ee52437f23783ed9baf4dfdd2120ca"} Nov 29 04:41:14 crc kubenswrapper[4799]: I1129 04:41:14.959368 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"c752b9e2f527d3036bd3401d3fb74dc2e4581efcbc8ef21b66560192d53a8db5"} Nov 29 04:41:14 crc kubenswrapper[4799]: I1129 04:41:14.959398 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"e2554ae4d2fa5616ffaebe57209155b976dd6b95a075335121ac508759ed2840"} Nov 29 04:41:14 crc kubenswrapper[4799]: I1129 04:41:14.959730 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 04:41:14 crc kubenswrapper[4799]: I1129 04:41:14.961361 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-crfwf" event={"ID":"0efffb4f-375f-49b8-8ff8-cc4149f71847","Type":"ContainerStarted","Data":"03690186bb3e965c79e3614f8e60e76154e5c4dbe53bdc256f9d8f4d576a7399"} Nov 29 04:41:15 crc kubenswrapper[4799]: I1129 04:41:15.057968 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=6.057932865 podStartE2EDuration="6.057932865s" podCreationTimestamp="2025-11-29 04:41:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:41:15.019952663 +0000 UTC m=+150.662883063" watchObservedRunningTime="2025-11-29 04:41:15.057932865 +0000 UTC m=+150.700863265" Nov 29 04:41:15 crc kubenswrapper[4799]: I1129 04:41:15.969880 4799 generic.go:334] "Generic (PLEG): container finished" podID="2544bf9f-2bb8-4661-997c-9b25a96103d4" containerID="9f6fff8b5ce57b269a0dd398202181c3a76dcc3aa7486e709184bf12d934965e" exitCode=0 Nov 29 04:41:15 crc kubenswrapper[4799]: I1129 04:41:15.969976 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jlvpn" event={"ID":"2544bf9f-2bb8-4661-997c-9b25a96103d4","Type":"ContainerDied","Data":"9f6fff8b5ce57b269a0dd398202181c3a76dcc3aa7486e709184bf12d934965e"} Nov 29 04:41:15 crc kubenswrapper[4799]: I1129 04:41:15.972578 4799 generic.go:334] "Generic (PLEG): container finished" podID="b0bccaa4-e99f-47b1-bf20-2895f289ffd7" containerID="cf73277195389762deb95a92e3bc34711b22d193b6a676db03a9fc1a80c521b2" exitCode=0 Nov 29 04:41:15 crc kubenswrapper[4799]: I1129 04:41:15.972641 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vl566" event={"ID":"b0bccaa4-e99f-47b1-bf20-2895f289ffd7","Type":"ContainerDied","Data":"cf73277195389762deb95a92e3bc34711b22d193b6a676db03a9fc1a80c521b2"} Nov 29 04:41:15 crc kubenswrapper[4799]: I1129 04:41:15.975483 4799 generic.go:334] "Generic (PLEG): container finished" podID="0efffb4f-375f-49b8-8ff8-cc4149f71847" containerID="03690186bb3e965c79e3614f8e60e76154e5c4dbe53bdc256f9d8f4d576a7399" exitCode=0 Nov 29 04:41:15 crc kubenswrapper[4799]: I1129 04:41:15.975572 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-crfwf" event={"ID":"0efffb4f-375f-49b8-8ff8-cc4149f71847","Type":"ContainerDied","Data":"03690186bb3e965c79e3614f8e60e76154e5c4dbe53bdc256f9d8f4d576a7399"} Nov 29 04:41:18 crc kubenswrapper[4799]: I1129 04:41:18.895983 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xlxlr"] Nov 29 04:41:18 crc kubenswrapper[4799]: I1129 04:41:18.896845 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xlxlr" podUID="ecc63394-2cf6-4a33-8e1b-a02042b8da4f" containerName="registry-server" containerID="cri-o://3ada1d3a01375e70ad1f7a656e7a48bc41707b0d852cddb11f2ce31cdfc6cb31" gracePeriod=2 Nov 29 04:41:19 crc kubenswrapper[4799]: I1129 04:41:19.000275 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-br2p6" event={"ID":"390f5da3-eb7e-48af-bcc0-666c9172403e","Type":"ContainerStarted","Data":"d33a9c8f1300805eea0ebb98dbb69b5b8e6f188ed0ac37e0ce2c483a4c223675"} Nov 29 04:41:19 crc kubenswrapper[4799]: I1129 04:41:19.005376 4799 generic.go:334] "Generic (PLEG): container finished" podID="0a4df431-deee-45d6-9c07-7a1616a23d91" containerID="49a1c143a43a03d721de97043194fb50472a26cb8e36f0f35330fe7a8895f5a9" exitCode=0 Nov 29 04:41:19 crc kubenswrapper[4799]: I1129 04:41:19.005457 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q6m25" event={"ID":"0a4df431-deee-45d6-9c07-7a1616a23d91","Type":"ContainerDied","Data":"49a1c143a43a03d721de97043194fb50472a26cb8e36f0f35330fe7a8895f5a9"} Nov 29 04:41:19 crc kubenswrapper[4799]: I1129 04:41:19.008234 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-crfwf" event={"ID":"0efffb4f-375f-49b8-8ff8-cc4149f71847","Type":"ContainerStarted","Data":"a89afd8e241176bf5408200e594f1ff14ba1b75edbbf64d7ca3f03bc4feb4eb2"} Nov 29 04:41:19 crc kubenswrapper[4799]: I1129 04:41:19.051723 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-br2p6" podStartSLOduration=5.93500915 podStartE2EDuration="56.051700825s" podCreationTimestamp="2025-11-29 04:40:23 +0000 UTC" firstStartedPulling="2025-11-29 04:40:25.391765646 +0000 UTC m=+101.034696056" lastFinishedPulling="2025-11-29 04:41:15.508457331 +0000 UTC m=+151.151387731" observedRunningTime="2025-11-29 04:41:19.032112138 +0000 UTC m=+154.675042538" watchObservedRunningTime="2025-11-29 04:41:19.051700825 +0000 UTC m=+154.694631225" Nov 29 04:41:19 crc kubenswrapper[4799]: I1129 04:41:19.073515 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-crfwf" podStartSLOduration=4.993459225 podStartE2EDuration="56.073490375s" podCreationTimestamp="2025-11-29 04:40:23 +0000 UTC" firstStartedPulling="2025-11-29 04:40:24.359739401 +0000 UTC m=+100.002669801" lastFinishedPulling="2025-11-29 04:41:15.439770551 +0000 UTC m=+151.082700951" observedRunningTime="2025-11-29 04:41:19.052254302 +0000 UTC m=+154.695184712" watchObservedRunningTime="2025-11-29 04:41:19.073490375 +0000 UTC m=+154.716420775" Nov 29 04:41:20 crc kubenswrapper[4799]: I1129 04:41:20.017635 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mplj2" event={"ID":"8a2cc255-bc5f-4253-bc2b-e74817971afe","Type":"ContainerStarted","Data":"1c19afd5be4b726bb334282559cead6c35f007ededfec1ac6fb204e8ee44c586"} Nov 29 04:41:20 crc kubenswrapper[4799]: I1129 04:41:20.022834 4799 generic.go:334] "Generic (PLEG): container finished" podID="ecc63394-2cf6-4a33-8e1b-a02042b8da4f" containerID="3ada1d3a01375e70ad1f7a656e7a48bc41707b0d852cddb11f2ce31cdfc6cb31" exitCode=0 Nov 29 04:41:20 crc kubenswrapper[4799]: I1129 04:41:20.022926 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xlxlr" event={"ID":"ecc63394-2cf6-4a33-8e1b-a02042b8da4f","Type":"ContainerDied","Data":"3ada1d3a01375e70ad1f7a656e7a48bc41707b0d852cddb11f2ce31cdfc6cb31"} Nov 29 04:41:21 crc kubenswrapper[4799]: I1129 04:41:21.031413 4799 generic.go:334] "Generic (PLEG): container finished" podID="8a2cc255-bc5f-4253-bc2b-e74817971afe" containerID="1c19afd5be4b726bb334282559cead6c35f007ededfec1ac6fb204e8ee44c586" exitCode=0 Nov 29 04:41:21 crc kubenswrapper[4799]: I1129 04:41:21.031491 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mplj2" event={"ID":"8a2cc255-bc5f-4253-bc2b-e74817971afe","Type":"ContainerDied","Data":"1c19afd5be4b726bb334282559cead6c35f007ededfec1ac6fb204e8ee44c586"} Nov 29 04:41:21 crc kubenswrapper[4799]: I1129 04:41:21.040781 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jlvpn" event={"ID":"2544bf9f-2bb8-4661-997c-9b25a96103d4","Type":"ContainerStarted","Data":"65ddff9b858f131e61390635199130e34b293bad0972d10e2d625b6a83beac00"} Nov 29 04:41:21 crc kubenswrapper[4799]: I1129 04:41:21.042963 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dnz56" event={"ID":"8734a259-eaa8-459e-b38d-e821e336a863","Type":"ContainerStarted","Data":"583b97c0ef0a8b6a73ba2cb895701337abd14f0f1c1188cb916a0b1ad750e5a6"} Nov 29 04:41:21 crc kubenswrapper[4799]: I1129 04:41:21.602023 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xlxlr" Nov 29 04:41:21 crc kubenswrapper[4799]: I1129 04:41:21.630705 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-smhtx\" (UniqueName: \"kubernetes.io/projected/ecc63394-2cf6-4a33-8e1b-a02042b8da4f-kube-api-access-smhtx\") pod \"ecc63394-2cf6-4a33-8e1b-a02042b8da4f\" (UID: \"ecc63394-2cf6-4a33-8e1b-a02042b8da4f\") " Nov 29 04:41:21 crc kubenswrapper[4799]: I1129 04:41:21.630821 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecc63394-2cf6-4a33-8e1b-a02042b8da4f-catalog-content\") pod \"ecc63394-2cf6-4a33-8e1b-a02042b8da4f\" (UID: \"ecc63394-2cf6-4a33-8e1b-a02042b8da4f\") " Nov 29 04:41:21 crc kubenswrapper[4799]: I1129 04:41:21.630907 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecc63394-2cf6-4a33-8e1b-a02042b8da4f-utilities\") pod \"ecc63394-2cf6-4a33-8e1b-a02042b8da4f\" (UID: \"ecc63394-2cf6-4a33-8e1b-a02042b8da4f\") " Nov 29 04:41:21 crc kubenswrapper[4799]: I1129 04:41:21.634207 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ecc63394-2cf6-4a33-8e1b-a02042b8da4f-utilities" (OuterVolumeSpecName: "utilities") pod "ecc63394-2cf6-4a33-8e1b-a02042b8da4f" (UID: "ecc63394-2cf6-4a33-8e1b-a02042b8da4f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:41:21 crc kubenswrapper[4799]: I1129 04:41:21.641808 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecc63394-2cf6-4a33-8e1b-a02042b8da4f-kube-api-access-smhtx" (OuterVolumeSpecName: "kube-api-access-smhtx") pod "ecc63394-2cf6-4a33-8e1b-a02042b8da4f" (UID: "ecc63394-2cf6-4a33-8e1b-a02042b8da4f"). InnerVolumeSpecName "kube-api-access-smhtx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:41:21 crc kubenswrapper[4799]: I1129 04:41:21.733942 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-smhtx\" (UniqueName: \"kubernetes.io/projected/ecc63394-2cf6-4a33-8e1b-a02042b8da4f-kube-api-access-smhtx\") on node \"crc\" DevicePath \"\"" Nov 29 04:41:21 crc kubenswrapper[4799]: I1129 04:41:21.733976 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecc63394-2cf6-4a33-8e1b-a02042b8da4f-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 04:41:21 crc kubenswrapper[4799]: I1129 04:41:21.745577 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ecc63394-2cf6-4a33-8e1b-a02042b8da4f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ecc63394-2cf6-4a33-8e1b-a02042b8da4f" (UID: "ecc63394-2cf6-4a33-8e1b-a02042b8da4f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:41:21 crc kubenswrapper[4799]: I1129 04:41:21.835710 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecc63394-2cf6-4a33-8e1b-a02042b8da4f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 04:41:22 crc kubenswrapper[4799]: I1129 04:41:22.054968 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xlxlr" event={"ID":"ecc63394-2cf6-4a33-8e1b-a02042b8da4f","Type":"ContainerDied","Data":"0aaa9681602c482a11c0da6b44cdb470ea19003b6981838a973b0bf11877fd1a"} Nov 29 04:41:22 crc kubenswrapper[4799]: I1129 04:41:22.055079 4799 scope.go:117] "RemoveContainer" containerID="3ada1d3a01375e70ad1f7a656e7a48bc41707b0d852cddb11f2ce31cdfc6cb31" Nov 29 04:41:22 crc kubenswrapper[4799]: I1129 04:41:22.054999 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xlxlr" Nov 29 04:41:22 crc kubenswrapper[4799]: I1129 04:41:22.057973 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mplj2" event={"ID":"8a2cc255-bc5f-4253-bc2b-e74817971afe","Type":"ContainerStarted","Data":"45e4fd6560df3f691aae7b1f960264d8d9f42d3b5c73eed74de71f06bde0bc21"} Nov 29 04:41:22 crc kubenswrapper[4799]: I1129 04:41:22.062002 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vl566" event={"ID":"b0bccaa4-e99f-47b1-bf20-2895f289ffd7","Type":"ContainerStarted","Data":"03dd3900fb4cd09590cdf95e591712a29201592633fc220927a9cda4e1806f23"} Nov 29 04:41:22 crc kubenswrapper[4799]: I1129 04:41:22.075428 4799 scope.go:117] "RemoveContainer" containerID="0bd9fb54a0c5c77b88fc6c0a563c36893388c12864861531d5ec4ae408caef94" Nov 29 04:41:22 crc kubenswrapper[4799]: I1129 04:41:22.089655 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jlvpn" podStartSLOduration=5.11520624 podStartE2EDuration="1m2.089630602s" podCreationTimestamp="2025-11-29 04:40:20 +0000 UTC" firstStartedPulling="2025-11-29 04:40:22.263923062 +0000 UTC m=+97.906853462" lastFinishedPulling="2025-11-29 04:41:19.238347424 +0000 UTC m=+154.881277824" observedRunningTime="2025-11-29 04:41:22.088698566 +0000 UTC m=+157.731628966" watchObservedRunningTime="2025-11-29 04:41:22.089630602 +0000 UTC m=+157.732561002" Nov 29 04:41:22 crc kubenswrapper[4799]: I1129 04:41:22.111121 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dnz56" podStartSLOduration=5.165474629 podStartE2EDuration="1m2.111098772s" podCreationTimestamp="2025-11-29 04:40:20 +0000 UTC" firstStartedPulling="2025-11-29 04:40:22.280118979 +0000 UTC m=+97.923049379" lastFinishedPulling="2025-11-29 04:41:19.225743122 +0000 UTC m=+154.868673522" observedRunningTime="2025-11-29 04:41:22.110628709 +0000 UTC m=+157.753559099" watchObservedRunningTime="2025-11-29 04:41:22.111098772 +0000 UTC m=+157.754029172" Nov 29 04:41:22 crc kubenswrapper[4799]: I1129 04:41:22.118853 4799 scope.go:117] "RemoveContainer" containerID="63c5e2dbdef44c5a12eef3bd8d44a34373e246e56f182fdd9c84c1823c5c0dc3" Nov 29 04:41:22 crc kubenswrapper[4799]: I1129 04:41:22.128996 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xlxlr"] Nov 29 04:41:22 crc kubenswrapper[4799]: I1129 04:41:22.140315 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xlxlr"] Nov 29 04:41:22 crc kubenswrapper[4799]: I1129 04:41:22.158329 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vl566" podStartSLOduration=1.937105684 podStartE2EDuration="1m1.158307752s" podCreationTimestamp="2025-11-29 04:40:21 +0000 UTC" firstStartedPulling="2025-11-29 04:40:22.258887022 +0000 UTC m=+97.901817422" lastFinishedPulling="2025-11-29 04:41:21.48008909 +0000 UTC m=+157.123019490" observedRunningTime="2025-11-29 04:41:22.154118585 +0000 UTC m=+157.797048995" watchObservedRunningTime="2025-11-29 04:41:22.158307752 +0000 UTC m=+157.801238152" Nov 29 04:41:22 crc kubenswrapper[4799]: I1129 04:41:22.666360 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ecc63394-2cf6-4a33-8e1b-a02042b8da4f" path="/var/lib/kubelet/pods/ecc63394-2cf6-4a33-8e1b-a02042b8da4f/volumes" Nov 29 04:41:23 crc kubenswrapper[4799]: I1129 04:41:23.071469 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q6m25" event={"ID":"0a4df431-deee-45d6-9c07-7a1616a23d91","Type":"ContainerStarted","Data":"32979ecb7f9458352b758f3316566a687c8ca038c4f6cb66ccacf500f508bf82"} Nov 29 04:41:23 crc kubenswrapper[4799]: I1129 04:41:23.119362 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mplj2" podStartSLOduration=2.558974227 podStartE2EDuration="1m2.119326811s" podCreationTimestamp="2025-11-29 04:40:21 +0000 UTC" firstStartedPulling="2025-11-29 04:40:22.256054519 +0000 UTC m=+97.898984919" lastFinishedPulling="2025-11-29 04:41:21.816407103 +0000 UTC m=+157.459337503" observedRunningTime="2025-11-29 04:41:23.115235516 +0000 UTC m=+158.758165916" watchObservedRunningTime="2025-11-29 04:41:23.119326811 +0000 UTC m=+158.762257211" Nov 29 04:41:23 crc kubenswrapper[4799]: I1129 04:41:23.154223 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-q6m25" podStartSLOduration=3.477785569 podStartE2EDuration="1m1.154196836s" podCreationTimestamp="2025-11-29 04:40:22 +0000 UTC" firstStartedPulling="2025-11-29 04:40:24.359997177 +0000 UTC m=+100.002927577" lastFinishedPulling="2025-11-29 04:41:22.036408444 +0000 UTC m=+157.679338844" observedRunningTime="2025-11-29 04:41:23.149210507 +0000 UTC m=+158.792140927" watchObservedRunningTime="2025-11-29 04:41:23.154196836 +0000 UTC m=+158.797127236" Nov 29 04:41:23 crc kubenswrapper[4799]: I1129 04:41:23.272091 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-q6m25" Nov 29 04:41:23 crc kubenswrapper[4799]: I1129 04:41:23.272154 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-q6m25" Nov 29 04:41:23 crc kubenswrapper[4799]: I1129 04:41:23.644127 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-crfwf" Nov 29 04:41:23 crc kubenswrapper[4799]: I1129 04:41:23.644219 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-crfwf" Nov 29 04:41:23 crc kubenswrapper[4799]: I1129 04:41:23.686171 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-crfwf" Nov 29 04:41:24 crc kubenswrapper[4799]: I1129 04:41:24.137561 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-crfwf" Nov 29 04:41:24 crc kubenswrapper[4799]: I1129 04:41:24.256530 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-br2p6" Nov 29 04:41:24 crc kubenswrapper[4799]: I1129 04:41:24.256588 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-br2p6" Nov 29 04:41:24 crc kubenswrapper[4799]: I1129 04:41:24.315177 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-q6m25" podUID="0a4df431-deee-45d6-9c07-7a1616a23d91" containerName="registry-server" probeResult="failure" output=< Nov 29 04:41:24 crc kubenswrapper[4799]: timeout: failed to connect service ":50051" within 1s Nov 29 04:41:24 crc kubenswrapper[4799]: > Nov 29 04:41:25 crc kubenswrapper[4799]: I1129 04:41:25.094447 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-crfwf"] Nov 29 04:41:25 crc kubenswrapper[4799]: I1129 04:41:25.298657 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-br2p6" podUID="390f5da3-eb7e-48af-bcc0-666c9172403e" containerName="registry-server" probeResult="failure" output=< Nov 29 04:41:25 crc kubenswrapper[4799]: timeout: failed to connect service ":50051" within 1s Nov 29 04:41:25 crc kubenswrapper[4799]: > Nov 29 04:41:26 crc kubenswrapper[4799]: I1129 04:41:26.089707 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-crfwf" podUID="0efffb4f-375f-49b8-8ff8-cc4149f71847" containerName="registry-server" containerID="cri-o://a89afd8e241176bf5408200e594f1ff14ba1b75edbbf64d7ca3f03bc4feb4eb2" gracePeriod=2 Nov 29 04:41:31 crc kubenswrapper[4799]: I1129 04:41:31.046994 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jlvpn" Nov 29 04:41:31 crc kubenswrapper[4799]: I1129 04:41:31.048096 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jlvpn" Nov 29 04:41:31 crc kubenswrapper[4799]: I1129 04:41:31.108607 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jlvpn" Nov 29 04:41:31 crc kubenswrapper[4799]: I1129 04:41:31.284084 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dnz56" Nov 29 04:41:31 crc kubenswrapper[4799]: I1129 04:41:31.284177 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dnz56" Nov 29 04:41:31 crc kubenswrapper[4799]: I1129 04:41:31.350808 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dnz56" Nov 29 04:41:31 crc kubenswrapper[4799]: I1129 04:41:31.541035 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mplj2" Nov 29 04:41:31 crc kubenswrapper[4799]: I1129 04:41:31.541454 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mplj2" Nov 29 04:41:31 crc kubenswrapper[4799]: I1129 04:41:31.605835 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mplj2" Nov 29 04:41:31 crc kubenswrapper[4799]: I1129 04:41:31.722594 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vl566" Nov 29 04:41:31 crc kubenswrapper[4799]: I1129 04:41:31.722724 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vl566" Nov 29 04:41:31 crc kubenswrapper[4799]: I1129 04:41:31.801479 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vl566" Nov 29 04:41:32 crc kubenswrapper[4799]: I1129 04:41:32.836231 4799 generic.go:334] "Generic (PLEG): container finished" podID="0efffb4f-375f-49b8-8ff8-cc4149f71847" containerID="a89afd8e241176bf5408200e594f1ff14ba1b75edbbf64d7ca3f03bc4feb4eb2" exitCode=0 Nov 29 04:41:32 crc kubenswrapper[4799]: I1129 04:41:32.836327 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-crfwf" event={"ID":"0efffb4f-375f-49b8-8ff8-cc4149f71847","Type":"ContainerDied","Data":"a89afd8e241176bf5408200e594f1ff14ba1b75edbbf64d7ca3f03bc4feb4eb2"} Nov 29 04:41:32 crc kubenswrapper[4799]: I1129 04:41:32.856718 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-crfwf" Nov 29 04:41:33 crc kubenswrapper[4799]: I1129 04:41:33.012166 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzsdx\" (UniqueName: \"kubernetes.io/projected/0efffb4f-375f-49b8-8ff8-cc4149f71847-kube-api-access-pzsdx\") pod \"0efffb4f-375f-49b8-8ff8-cc4149f71847\" (UID: \"0efffb4f-375f-49b8-8ff8-cc4149f71847\") " Nov 29 04:41:33 crc kubenswrapper[4799]: I1129 04:41:33.013340 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0efffb4f-375f-49b8-8ff8-cc4149f71847-catalog-content\") pod \"0efffb4f-375f-49b8-8ff8-cc4149f71847\" (UID: \"0efffb4f-375f-49b8-8ff8-cc4149f71847\") " Nov 29 04:41:33 crc kubenswrapper[4799]: I1129 04:41:33.013525 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0efffb4f-375f-49b8-8ff8-cc4149f71847-utilities\") pod \"0efffb4f-375f-49b8-8ff8-cc4149f71847\" (UID: \"0efffb4f-375f-49b8-8ff8-cc4149f71847\") " Nov 29 04:41:33 crc kubenswrapper[4799]: I1129 04:41:33.015072 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0efffb4f-375f-49b8-8ff8-cc4149f71847-utilities" (OuterVolumeSpecName: "utilities") pod "0efffb4f-375f-49b8-8ff8-cc4149f71847" (UID: "0efffb4f-375f-49b8-8ff8-cc4149f71847"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:41:33 crc kubenswrapper[4799]: I1129 04:41:33.022838 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0efffb4f-375f-49b8-8ff8-cc4149f71847-kube-api-access-pzsdx" (OuterVolumeSpecName: "kube-api-access-pzsdx") pod "0efffb4f-375f-49b8-8ff8-cc4149f71847" (UID: "0efffb4f-375f-49b8-8ff8-cc4149f71847"). InnerVolumeSpecName "kube-api-access-pzsdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:41:33 crc kubenswrapper[4799]: I1129 04:41:33.044930 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0efffb4f-375f-49b8-8ff8-cc4149f71847-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0efffb4f-375f-49b8-8ff8-cc4149f71847" (UID: "0efffb4f-375f-49b8-8ff8-cc4149f71847"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:41:33 crc kubenswrapper[4799]: I1129 04:41:33.116199 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzsdx\" (UniqueName: \"kubernetes.io/projected/0efffb4f-375f-49b8-8ff8-cc4149f71847-kube-api-access-pzsdx\") on node \"crc\" DevicePath \"\"" Nov 29 04:41:33 crc kubenswrapper[4799]: I1129 04:41:33.116287 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0efffb4f-375f-49b8-8ff8-cc4149f71847-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 04:41:33 crc kubenswrapper[4799]: I1129 04:41:33.116306 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0efffb4f-375f-49b8-8ff8-cc4149f71847-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 04:41:33 crc kubenswrapper[4799]: I1129 04:41:33.349300 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-q6m25" Nov 29 04:41:33 crc kubenswrapper[4799]: I1129 04:41:33.418888 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-q6m25" Nov 29 04:41:33 crc kubenswrapper[4799]: I1129 04:41:33.848611 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-crfwf" event={"ID":"0efffb4f-375f-49b8-8ff8-cc4149f71847","Type":"ContainerDied","Data":"0ef24046dbe9b6f9e6c24d89f656148f1cb602813654ca6935f1b84d18824e24"} Nov 29 04:41:33 crc kubenswrapper[4799]: I1129 04:41:33.849053 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-crfwf" Nov 29 04:41:33 crc kubenswrapper[4799]: I1129 04:41:33.849186 4799 scope.go:117] "RemoveContainer" containerID="a89afd8e241176bf5408200e594f1ff14ba1b75edbbf64d7ca3f03bc4feb4eb2" Nov 29 04:41:33 crc kubenswrapper[4799]: I1129 04:41:33.910995 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-crfwf"] Nov 29 04:41:33 crc kubenswrapper[4799]: I1129 04:41:33.912746 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dnz56" Nov 29 04:41:33 crc kubenswrapper[4799]: I1129 04:41:33.917746 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-crfwf"] Nov 29 04:41:33 crc kubenswrapper[4799]: I1129 04:41:33.924648 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jlvpn" Nov 29 04:41:33 crc kubenswrapper[4799]: I1129 04:41:33.926544 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vl566" Nov 29 04:41:33 crc kubenswrapper[4799]: I1129 04:41:33.949726 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mplj2" Nov 29 04:41:34 crc kubenswrapper[4799]: I1129 04:41:34.311216 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-br2p6" Nov 29 04:41:34 crc kubenswrapper[4799]: I1129 04:41:34.361085 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-br2p6" Nov 29 04:41:34 crc kubenswrapper[4799]: I1129 04:41:34.430863 4799 scope.go:117] "RemoveContainer" containerID="03690186bb3e965c79e3614f8e60e76154e5c4dbe53bdc256f9d8f4d576a7399" Nov 29 04:41:34 crc kubenswrapper[4799]: I1129 04:41:34.475020 4799 scope.go:117] "RemoveContainer" containerID="9657fd890f3b8b97e14129e9f4ec54425883bd680d8da9fdad6333ffc22764dc" Nov 29 04:41:34 crc kubenswrapper[4799]: I1129 04:41:34.669851 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0efffb4f-375f-49b8-8ff8-cc4149f71847" path="/var/lib/kubelet/pods/0efffb4f-375f-49b8-8ff8-cc4149f71847/volumes" Nov 29 04:41:35 crc kubenswrapper[4799]: I1129 04:41:35.483867 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" podUID="bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d" containerName="oauth-openshift" containerID="cri-o://a6408959e5fbc68bed51a4dce9165a56d9c58cc0083289a8e606eb87e4511e8c" gracePeriod=15 Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.341755 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vl566"] Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.342508 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vl566" podUID="b0bccaa4-e99f-47b1-bf20-2895f289ffd7" containerName="registry-server" containerID="cri-o://03dd3900fb4cd09590cdf95e591712a29201592633fc220927a9cda4e1806f23" gracePeriod=2 Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.369879 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.482204 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-system-router-certs\") pod \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.482278 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-system-trusted-ca-bundle\") pod \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.482327 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-system-service-ca\") pod \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.482350 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-user-template-login\") pod \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.482371 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-audit-dir\") pod \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.482448 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-system-session\") pod \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.482477 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrk57\" (UniqueName: \"kubernetes.io/projected/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-kube-api-access-jrk57\") pod \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.482506 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-system-serving-cert\") pod \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.482524 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-user-idp-0-file-data\") pod \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.482544 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-system-ocp-branding-template\") pod \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.482565 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-system-cliconfig\") pod \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.482595 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-audit-policies\") pod \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.482610 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-user-template-error\") pod \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.482627 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-user-template-provider-selection\") pod \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\" (UID: \"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d\") " Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.487423 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d" (UID: "bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.487728 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d" (UID: "bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.487969 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d" (UID: "bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.488448 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d" (UID: "bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.491574 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d" (UID: "bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.525050 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-kube-api-access-jrk57" (OuterVolumeSpecName: "kube-api-access-jrk57") pod "bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d" (UID: "bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d"). InnerVolumeSpecName "kube-api-access-jrk57". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.537088 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d" (UID: "bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.537389 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d" (UID: "bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.538914 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d" (UID: "bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.541405 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d" (UID: "bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.541695 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d" (UID: "bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.542077 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d" (UID: "bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.554587 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d" (UID: "bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.555311 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d" (UID: "bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.559052 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mplj2"] Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.584209 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.584337 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrk57\" (UniqueName: \"kubernetes.io/projected/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-kube-api-access-jrk57\") on node \"crc\" DevicePath \"\"" Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.584351 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.584365 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.584381 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.584411 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.584424 4799 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.584436 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.584451 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.584488 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.584502 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.584514 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.584526 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.584539 4799 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.872573 4799 generic.go:334] "Generic (PLEG): container finished" podID="b0bccaa4-e99f-47b1-bf20-2895f289ffd7" containerID="03dd3900fb4cd09590cdf95e591712a29201592633fc220927a9cda4e1806f23" exitCode=0 Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.872646 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vl566" event={"ID":"b0bccaa4-e99f-47b1-bf20-2895f289ffd7","Type":"ContainerDied","Data":"03dd3900fb4cd09590cdf95e591712a29201592633fc220927a9cda4e1806f23"} Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.874863 4799 generic.go:334] "Generic (PLEG): container finished" podID="bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d" containerID="a6408959e5fbc68bed51a4dce9165a56d9c58cc0083289a8e606eb87e4511e8c" exitCode=0 Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.874904 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" event={"ID":"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d","Type":"ContainerDied","Data":"a6408959e5fbc68bed51a4dce9165a56d9c58cc0083289a8e606eb87e4511e8c"} Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.875010 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.875042 4799 scope.go:117] "RemoveContainer" containerID="a6408959e5fbc68bed51a4dce9165a56d9c58cc0083289a8e606eb87e4511e8c" Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.875024 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-vdxqp" event={"ID":"bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d","Type":"ContainerDied","Data":"1ad557578cb04dc2f7758cf89412842a705604f1a60338baf488e205cbcbb90a"} Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.875399 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mplj2" podUID="8a2cc255-bc5f-4253-bc2b-e74817971afe" containerName="registry-server" containerID="cri-o://45e4fd6560df3f691aae7b1f960264d8d9f42d3b5c73eed74de71f06bde0bc21" gracePeriod=2 Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.894344 4799 scope.go:117] "RemoveContainer" containerID="a6408959e5fbc68bed51a4dce9165a56d9c58cc0083289a8e606eb87e4511e8c" Nov 29 04:41:36 crc kubenswrapper[4799]: E1129 04:41:36.895035 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6408959e5fbc68bed51a4dce9165a56d9c58cc0083289a8e606eb87e4511e8c\": container with ID starting with a6408959e5fbc68bed51a4dce9165a56d9c58cc0083289a8e606eb87e4511e8c not found: ID does not exist" containerID="a6408959e5fbc68bed51a4dce9165a56d9c58cc0083289a8e606eb87e4511e8c" Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.895092 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6408959e5fbc68bed51a4dce9165a56d9c58cc0083289a8e606eb87e4511e8c"} err="failed to get container status \"a6408959e5fbc68bed51a4dce9165a56d9c58cc0083289a8e606eb87e4511e8c\": rpc error: code = NotFound desc = could not find container \"a6408959e5fbc68bed51a4dce9165a56d9c58cc0083289a8e606eb87e4511e8c\": container with ID starting with a6408959e5fbc68bed51a4dce9165a56d9c58cc0083289a8e606eb87e4511e8c not found: ID does not exist" Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.901772 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-vdxqp"] Nov 29 04:41:36 crc kubenswrapper[4799]: I1129 04:41:36.904418 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-vdxqp"] Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.191560 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-75c755456-6z6nn"] Nov 29 04:41:37 crc kubenswrapper[4799]: E1129 04:41:37.192313 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecc63394-2cf6-4a33-8e1b-a02042b8da4f" containerName="extract-utilities" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.192383 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecc63394-2cf6-4a33-8e1b-a02042b8da4f" containerName="extract-utilities" Nov 29 04:41:37 crc kubenswrapper[4799]: E1129 04:41:37.192484 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecc63394-2cf6-4a33-8e1b-a02042b8da4f" containerName="registry-server" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.192548 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecc63394-2cf6-4a33-8e1b-a02042b8da4f" containerName="registry-server" Nov 29 04:41:37 crc kubenswrapper[4799]: E1129 04:41:37.192606 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d" containerName="oauth-openshift" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.192969 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d" containerName="oauth-openshift" Nov 29 04:41:37 crc kubenswrapper[4799]: E1129 04:41:37.193040 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0efffb4f-375f-49b8-8ff8-cc4149f71847" containerName="extract-utilities" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.193092 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="0efffb4f-375f-49b8-8ff8-cc4149f71847" containerName="extract-utilities" Nov 29 04:41:37 crc kubenswrapper[4799]: E1129 04:41:37.193141 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecc63394-2cf6-4a33-8e1b-a02042b8da4f" containerName="extract-content" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.193198 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecc63394-2cf6-4a33-8e1b-a02042b8da4f" containerName="extract-content" Nov 29 04:41:37 crc kubenswrapper[4799]: E1129 04:41:37.193254 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0efffb4f-375f-49b8-8ff8-cc4149f71847" containerName="extract-content" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.193304 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="0efffb4f-375f-49b8-8ff8-cc4149f71847" containerName="extract-content" Nov 29 04:41:37 crc kubenswrapper[4799]: E1129 04:41:37.193371 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0efffb4f-375f-49b8-8ff8-cc4149f71847" containerName="registry-server" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.193423 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="0efffb4f-375f-49b8-8ff8-cc4149f71847" containerName="registry-server" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.193590 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="0efffb4f-375f-49b8-8ff8-cc4149f71847" containerName="registry-server" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.193672 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecc63394-2cf6-4a33-8e1b-a02042b8da4f" containerName="registry-server" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.193786 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d" containerName="oauth-openshift" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.194366 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-75c755456-6z6nn" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.196582 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.197634 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.200593 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.201321 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.201577 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.201659 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.202549 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.202755 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.202929 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.203071 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.205324 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.206205 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.212022 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.230143 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.230480 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.267952 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-75c755456-6z6nn"] Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.294237 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7ed02282-fade-4fad-b82e-cda665ec7998-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-75c755456-6z6nn\" (UID: \"7ed02282-fade-4fad-b82e-cda665ec7998\") " pod="openshift-authentication/oauth-openshift-75c755456-6z6nn" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.294622 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7ed02282-fade-4fad-b82e-cda665ec7998-v4-0-config-system-router-certs\") pod \"oauth-openshift-75c755456-6z6nn\" (UID: \"7ed02282-fade-4fad-b82e-cda665ec7998\") " pod="openshift-authentication/oauth-openshift-75c755456-6z6nn" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.294749 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7ed02282-fade-4fad-b82e-cda665ec7998-v4-0-config-system-session\") pod \"oauth-openshift-75c755456-6z6nn\" (UID: \"7ed02282-fade-4fad-b82e-cda665ec7998\") " pod="openshift-authentication/oauth-openshift-75c755456-6z6nn" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.294910 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7ed02282-fade-4fad-b82e-cda665ec7998-v4-0-config-system-serving-cert\") pod \"oauth-openshift-75c755456-6z6nn\" (UID: \"7ed02282-fade-4fad-b82e-cda665ec7998\") " pod="openshift-authentication/oauth-openshift-75c755456-6z6nn" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.295007 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7ed02282-fade-4fad-b82e-cda665ec7998-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-75c755456-6z6nn\" (UID: \"7ed02282-fade-4fad-b82e-cda665ec7998\") " pod="openshift-authentication/oauth-openshift-75c755456-6z6nn" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.295151 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7ed02282-fade-4fad-b82e-cda665ec7998-audit-policies\") pod \"oauth-openshift-75c755456-6z6nn\" (UID: \"7ed02282-fade-4fad-b82e-cda665ec7998\") " pod="openshift-authentication/oauth-openshift-75c755456-6z6nn" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.295262 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7ed02282-fade-4fad-b82e-cda665ec7998-v4-0-config-user-template-error\") pod \"oauth-openshift-75c755456-6z6nn\" (UID: \"7ed02282-fade-4fad-b82e-cda665ec7998\") " pod="openshift-authentication/oauth-openshift-75c755456-6z6nn" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.295367 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7ed02282-fade-4fad-b82e-cda665ec7998-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-75c755456-6z6nn\" (UID: \"7ed02282-fade-4fad-b82e-cda665ec7998\") " pod="openshift-authentication/oauth-openshift-75c755456-6z6nn" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.295654 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7ed02282-fade-4fad-b82e-cda665ec7998-audit-dir\") pod \"oauth-openshift-75c755456-6z6nn\" (UID: \"7ed02282-fade-4fad-b82e-cda665ec7998\") " pod="openshift-authentication/oauth-openshift-75c755456-6z6nn" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.295773 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jg5h6\" (UniqueName: \"kubernetes.io/projected/7ed02282-fade-4fad-b82e-cda665ec7998-kube-api-access-jg5h6\") pod \"oauth-openshift-75c755456-6z6nn\" (UID: \"7ed02282-fade-4fad-b82e-cda665ec7998\") " pod="openshift-authentication/oauth-openshift-75c755456-6z6nn" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.295918 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7ed02282-fade-4fad-b82e-cda665ec7998-v4-0-config-system-service-ca\") pod \"oauth-openshift-75c755456-6z6nn\" (UID: \"7ed02282-fade-4fad-b82e-cda665ec7998\") " pod="openshift-authentication/oauth-openshift-75c755456-6z6nn" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.296037 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7ed02282-fade-4fad-b82e-cda665ec7998-v4-0-config-system-cliconfig\") pod \"oauth-openshift-75c755456-6z6nn\" (UID: \"7ed02282-fade-4fad-b82e-cda665ec7998\") " pod="openshift-authentication/oauth-openshift-75c755456-6z6nn" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.296148 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7ed02282-fade-4fad-b82e-cda665ec7998-v4-0-config-user-template-login\") pod \"oauth-openshift-75c755456-6z6nn\" (UID: \"7ed02282-fade-4fad-b82e-cda665ec7998\") " pod="openshift-authentication/oauth-openshift-75c755456-6z6nn" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.296273 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7ed02282-fade-4fad-b82e-cda665ec7998-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-75c755456-6z6nn\" (UID: \"7ed02282-fade-4fad-b82e-cda665ec7998\") " pod="openshift-authentication/oauth-openshift-75c755456-6z6nn" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.397954 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7ed02282-fade-4fad-b82e-cda665ec7998-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-75c755456-6z6nn\" (UID: \"7ed02282-fade-4fad-b82e-cda665ec7998\") " pod="openshift-authentication/oauth-openshift-75c755456-6z6nn" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.398302 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7ed02282-fade-4fad-b82e-cda665ec7998-audit-dir\") pod \"oauth-openshift-75c755456-6z6nn\" (UID: \"7ed02282-fade-4fad-b82e-cda665ec7998\") " pod="openshift-authentication/oauth-openshift-75c755456-6z6nn" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.398414 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jg5h6\" (UniqueName: \"kubernetes.io/projected/7ed02282-fade-4fad-b82e-cda665ec7998-kube-api-access-jg5h6\") pod \"oauth-openshift-75c755456-6z6nn\" (UID: \"7ed02282-fade-4fad-b82e-cda665ec7998\") " pod="openshift-authentication/oauth-openshift-75c755456-6z6nn" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.398538 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7ed02282-fade-4fad-b82e-cda665ec7998-v4-0-config-system-cliconfig\") pod \"oauth-openshift-75c755456-6z6nn\" (UID: \"7ed02282-fade-4fad-b82e-cda665ec7998\") " pod="openshift-authentication/oauth-openshift-75c755456-6z6nn" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.398659 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7ed02282-fade-4fad-b82e-cda665ec7998-v4-0-config-system-service-ca\") pod \"oauth-openshift-75c755456-6z6nn\" (UID: \"7ed02282-fade-4fad-b82e-cda665ec7998\") " pod="openshift-authentication/oauth-openshift-75c755456-6z6nn" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.398782 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7ed02282-fade-4fad-b82e-cda665ec7998-v4-0-config-user-template-login\") pod \"oauth-openshift-75c755456-6z6nn\" (UID: \"7ed02282-fade-4fad-b82e-cda665ec7998\") " pod="openshift-authentication/oauth-openshift-75c755456-6z6nn" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.398935 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7ed02282-fade-4fad-b82e-cda665ec7998-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-75c755456-6z6nn\" (UID: \"7ed02282-fade-4fad-b82e-cda665ec7998\") " pod="openshift-authentication/oauth-openshift-75c755456-6z6nn" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.399064 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7ed02282-fade-4fad-b82e-cda665ec7998-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-75c755456-6z6nn\" (UID: \"7ed02282-fade-4fad-b82e-cda665ec7998\") " pod="openshift-authentication/oauth-openshift-75c755456-6z6nn" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.399176 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7ed02282-fade-4fad-b82e-cda665ec7998-v4-0-config-system-router-certs\") pod \"oauth-openshift-75c755456-6z6nn\" (UID: \"7ed02282-fade-4fad-b82e-cda665ec7998\") " pod="openshift-authentication/oauth-openshift-75c755456-6z6nn" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.399272 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7ed02282-fade-4fad-b82e-cda665ec7998-v4-0-config-system-session\") pod \"oauth-openshift-75c755456-6z6nn\" (UID: \"7ed02282-fade-4fad-b82e-cda665ec7998\") " pod="openshift-authentication/oauth-openshift-75c755456-6z6nn" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.399376 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7ed02282-fade-4fad-b82e-cda665ec7998-v4-0-config-system-serving-cert\") pod \"oauth-openshift-75c755456-6z6nn\" (UID: \"7ed02282-fade-4fad-b82e-cda665ec7998\") " pod="openshift-authentication/oauth-openshift-75c755456-6z6nn" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.399472 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7ed02282-fade-4fad-b82e-cda665ec7998-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-75c755456-6z6nn\" (UID: \"7ed02282-fade-4fad-b82e-cda665ec7998\") " pod="openshift-authentication/oauth-openshift-75c755456-6z6nn" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.399588 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7ed02282-fade-4fad-b82e-cda665ec7998-audit-policies\") pod \"oauth-openshift-75c755456-6z6nn\" (UID: \"7ed02282-fade-4fad-b82e-cda665ec7998\") " pod="openshift-authentication/oauth-openshift-75c755456-6z6nn" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.399731 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7ed02282-fade-4fad-b82e-cda665ec7998-v4-0-config-user-template-error\") pod \"oauth-openshift-75c755456-6z6nn\" (UID: \"7ed02282-fade-4fad-b82e-cda665ec7998\") " pod="openshift-authentication/oauth-openshift-75c755456-6z6nn" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.399478 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7ed02282-fade-4fad-b82e-cda665ec7998-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-75c755456-6z6nn\" (UID: \"7ed02282-fade-4fad-b82e-cda665ec7998\") " pod="openshift-authentication/oauth-openshift-75c755456-6z6nn" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.401545 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7ed02282-fade-4fad-b82e-cda665ec7998-audit-dir\") pod \"oauth-openshift-75c755456-6z6nn\" (UID: \"7ed02282-fade-4fad-b82e-cda665ec7998\") " pod="openshift-authentication/oauth-openshift-75c755456-6z6nn" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.401960 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7ed02282-fade-4fad-b82e-cda665ec7998-audit-policies\") pod \"oauth-openshift-75c755456-6z6nn\" (UID: \"7ed02282-fade-4fad-b82e-cda665ec7998\") " pod="openshift-authentication/oauth-openshift-75c755456-6z6nn" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.402897 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7ed02282-fade-4fad-b82e-cda665ec7998-v4-0-config-system-cliconfig\") pod \"oauth-openshift-75c755456-6z6nn\" (UID: \"7ed02282-fade-4fad-b82e-cda665ec7998\") " pod="openshift-authentication/oauth-openshift-75c755456-6z6nn" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.404401 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7ed02282-fade-4fad-b82e-cda665ec7998-v4-0-config-system-service-ca\") pod \"oauth-openshift-75c755456-6z6nn\" (UID: \"7ed02282-fade-4fad-b82e-cda665ec7998\") " pod="openshift-authentication/oauth-openshift-75c755456-6z6nn" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.404805 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7ed02282-fade-4fad-b82e-cda665ec7998-v4-0-config-user-template-error\") pod \"oauth-openshift-75c755456-6z6nn\" (UID: \"7ed02282-fade-4fad-b82e-cda665ec7998\") " pod="openshift-authentication/oauth-openshift-75c755456-6z6nn" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.405323 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7ed02282-fade-4fad-b82e-cda665ec7998-v4-0-config-system-session\") pod \"oauth-openshift-75c755456-6z6nn\" (UID: \"7ed02282-fade-4fad-b82e-cda665ec7998\") " pod="openshift-authentication/oauth-openshift-75c755456-6z6nn" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.405460 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7ed02282-fade-4fad-b82e-cda665ec7998-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-75c755456-6z6nn\" (UID: \"7ed02282-fade-4fad-b82e-cda665ec7998\") " pod="openshift-authentication/oauth-openshift-75c755456-6z6nn" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.407467 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7ed02282-fade-4fad-b82e-cda665ec7998-v4-0-config-system-router-certs\") pod \"oauth-openshift-75c755456-6z6nn\" (UID: \"7ed02282-fade-4fad-b82e-cda665ec7998\") " pod="openshift-authentication/oauth-openshift-75c755456-6z6nn" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.407918 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7ed02282-fade-4fad-b82e-cda665ec7998-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-75c755456-6z6nn\" (UID: \"7ed02282-fade-4fad-b82e-cda665ec7998\") " pod="openshift-authentication/oauth-openshift-75c755456-6z6nn" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.409004 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7ed02282-fade-4fad-b82e-cda665ec7998-v4-0-config-user-template-login\") pod \"oauth-openshift-75c755456-6z6nn\" (UID: \"7ed02282-fade-4fad-b82e-cda665ec7998\") " pod="openshift-authentication/oauth-openshift-75c755456-6z6nn" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.409283 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7ed02282-fade-4fad-b82e-cda665ec7998-v4-0-config-system-serving-cert\") pod \"oauth-openshift-75c755456-6z6nn\" (UID: \"7ed02282-fade-4fad-b82e-cda665ec7998\") " pod="openshift-authentication/oauth-openshift-75c755456-6z6nn" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.416131 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7ed02282-fade-4fad-b82e-cda665ec7998-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-75c755456-6z6nn\" (UID: \"7ed02282-fade-4fad-b82e-cda665ec7998\") " pod="openshift-authentication/oauth-openshift-75c755456-6z6nn" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.417863 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jg5h6\" (UniqueName: \"kubernetes.io/projected/7ed02282-fade-4fad-b82e-cda665ec7998-kube-api-access-jg5h6\") pod \"oauth-openshift-75c755456-6z6nn\" (UID: \"7ed02282-fade-4fad-b82e-cda665ec7998\") " pod="openshift-authentication/oauth-openshift-75c755456-6z6nn" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.498289 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.498766 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.507981 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-75c755456-6z6nn" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.709836 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vl566" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.778781 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mplj2" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.785778 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-75c755456-6z6nn"] Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.806311 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xvpw\" (UniqueName: \"kubernetes.io/projected/8a2cc255-bc5f-4253-bc2b-e74817971afe-kube-api-access-7xvpw\") pod \"8a2cc255-bc5f-4253-bc2b-e74817971afe\" (UID: \"8a2cc255-bc5f-4253-bc2b-e74817971afe\") " Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.806440 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0bccaa4-e99f-47b1-bf20-2895f289ffd7-utilities\") pod \"b0bccaa4-e99f-47b1-bf20-2895f289ffd7\" (UID: \"b0bccaa4-e99f-47b1-bf20-2895f289ffd7\") " Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.806534 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0bccaa4-e99f-47b1-bf20-2895f289ffd7-catalog-content\") pod \"b0bccaa4-e99f-47b1-bf20-2895f289ffd7\" (UID: \"b0bccaa4-e99f-47b1-bf20-2895f289ffd7\") " Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.806682 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a2cc255-bc5f-4253-bc2b-e74817971afe-catalog-content\") pod \"8a2cc255-bc5f-4253-bc2b-e74817971afe\" (UID: \"8a2cc255-bc5f-4253-bc2b-e74817971afe\") " Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.806768 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a2cc255-bc5f-4253-bc2b-e74817971afe-utilities\") pod \"8a2cc255-bc5f-4253-bc2b-e74817971afe\" (UID: \"8a2cc255-bc5f-4253-bc2b-e74817971afe\") " Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.806841 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d676x\" (UniqueName: \"kubernetes.io/projected/b0bccaa4-e99f-47b1-bf20-2895f289ffd7-kube-api-access-d676x\") pod \"b0bccaa4-e99f-47b1-bf20-2895f289ffd7\" (UID: \"b0bccaa4-e99f-47b1-bf20-2895f289ffd7\") " Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.809275 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0bccaa4-e99f-47b1-bf20-2895f289ffd7-utilities" (OuterVolumeSpecName: "utilities") pod "b0bccaa4-e99f-47b1-bf20-2895f289ffd7" (UID: "b0bccaa4-e99f-47b1-bf20-2895f289ffd7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.809484 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a2cc255-bc5f-4253-bc2b-e74817971afe-utilities" (OuterVolumeSpecName: "utilities") pod "8a2cc255-bc5f-4253-bc2b-e74817971afe" (UID: "8a2cc255-bc5f-4253-bc2b-e74817971afe"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.812395 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0bccaa4-e99f-47b1-bf20-2895f289ffd7-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.812434 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a2cc255-bc5f-4253-bc2b-e74817971afe-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.814083 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a2cc255-bc5f-4253-bc2b-e74817971afe-kube-api-access-7xvpw" (OuterVolumeSpecName: "kube-api-access-7xvpw") pod "8a2cc255-bc5f-4253-bc2b-e74817971afe" (UID: "8a2cc255-bc5f-4253-bc2b-e74817971afe"). InnerVolumeSpecName "kube-api-access-7xvpw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.816225 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0bccaa4-e99f-47b1-bf20-2895f289ffd7-kube-api-access-d676x" (OuterVolumeSpecName: "kube-api-access-d676x") pod "b0bccaa4-e99f-47b1-bf20-2895f289ffd7" (UID: "b0bccaa4-e99f-47b1-bf20-2895f289ffd7"). InnerVolumeSpecName "kube-api-access-d676x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.862358 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a2cc255-bc5f-4253-bc2b-e74817971afe-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8a2cc255-bc5f-4253-bc2b-e74817971afe" (UID: "8a2cc255-bc5f-4253-bc2b-e74817971afe"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.872368 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0bccaa4-e99f-47b1-bf20-2895f289ffd7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b0bccaa4-e99f-47b1-bf20-2895f289ffd7" (UID: "b0bccaa4-e99f-47b1-bf20-2895f289ffd7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.886053 4799 generic.go:334] "Generic (PLEG): container finished" podID="8a2cc255-bc5f-4253-bc2b-e74817971afe" containerID="45e4fd6560df3f691aae7b1f960264d8d9f42d3b5c73eed74de71f06bde0bc21" exitCode=0 Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.886149 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mplj2" event={"ID":"8a2cc255-bc5f-4253-bc2b-e74817971afe","Type":"ContainerDied","Data":"45e4fd6560df3f691aae7b1f960264d8d9f42d3b5c73eed74de71f06bde0bc21"} Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.886278 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mplj2" event={"ID":"8a2cc255-bc5f-4253-bc2b-e74817971afe","Type":"ContainerDied","Data":"e25cd57ab8d2aa6363166cd416b592c66ebd00c0e1f196b0aeee28f1b480b110"} Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.886190 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mplj2" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.886317 4799 scope.go:117] "RemoveContainer" containerID="45e4fd6560df3f691aae7b1f960264d8d9f42d3b5c73eed74de71f06bde0bc21" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.891689 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vl566" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.891986 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vl566" event={"ID":"b0bccaa4-e99f-47b1-bf20-2895f289ffd7","Type":"ContainerDied","Data":"f6237c5e1897b877aa2f5608a2b48444b4d810c1670ad3a7cd116775cf533bfe"} Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.895462 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-75c755456-6z6nn" event={"ID":"7ed02282-fade-4fad-b82e-cda665ec7998","Type":"ContainerStarted","Data":"73ef759f7af2a33523e9d085d01b912e69c5f6e7076c556ae77100d2d71087ad"} Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.905537 4799 scope.go:117] "RemoveContainer" containerID="1c19afd5be4b726bb334282559cead6c35f007ededfec1ac6fb204e8ee44c586" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.913355 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7xvpw\" (UniqueName: \"kubernetes.io/projected/8a2cc255-bc5f-4253-bc2b-e74817971afe-kube-api-access-7xvpw\") on node \"crc\" DevicePath \"\"" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.913389 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0bccaa4-e99f-47b1-bf20-2895f289ffd7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.913402 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a2cc255-bc5f-4253-bc2b-e74817971afe-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.913413 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d676x\" (UniqueName: \"kubernetes.io/projected/b0bccaa4-e99f-47b1-bf20-2895f289ffd7-kube-api-access-d676x\") on node \"crc\" DevicePath \"\"" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.928491 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mplj2"] Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.945728 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mplj2"] Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.952830 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vl566"] Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.956144 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vl566"] Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.956413 4799 scope.go:117] "RemoveContainer" containerID="c4cf32469a3d4feffd40673f38d4a39abb101d5196c09641054e3eff516aa22e" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.972370 4799 scope.go:117] "RemoveContainer" containerID="45e4fd6560df3f691aae7b1f960264d8d9f42d3b5c73eed74de71f06bde0bc21" Nov 29 04:41:37 crc kubenswrapper[4799]: E1129 04:41:37.972873 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45e4fd6560df3f691aae7b1f960264d8d9f42d3b5c73eed74de71f06bde0bc21\": container with ID starting with 45e4fd6560df3f691aae7b1f960264d8d9f42d3b5c73eed74de71f06bde0bc21 not found: ID does not exist" containerID="45e4fd6560df3f691aae7b1f960264d8d9f42d3b5c73eed74de71f06bde0bc21" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.972920 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45e4fd6560df3f691aae7b1f960264d8d9f42d3b5c73eed74de71f06bde0bc21"} err="failed to get container status \"45e4fd6560df3f691aae7b1f960264d8d9f42d3b5c73eed74de71f06bde0bc21\": rpc error: code = NotFound desc = could not find container \"45e4fd6560df3f691aae7b1f960264d8d9f42d3b5c73eed74de71f06bde0bc21\": container with ID starting with 45e4fd6560df3f691aae7b1f960264d8d9f42d3b5c73eed74de71f06bde0bc21 not found: ID does not exist" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.972951 4799 scope.go:117] "RemoveContainer" containerID="1c19afd5be4b726bb334282559cead6c35f007ededfec1ac6fb204e8ee44c586" Nov 29 04:41:37 crc kubenswrapper[4799]: E1129 04:41:37.973461 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c19afd5be4b726bb334282559cead6c35f007ededfec1ac6fb204e8ee44c586\": container with ID starting with 1c19afd5be4b726bb334282559cead6c35f007ededfec1ac6fb204e8ee44c586 not found: ID does not exist" containerID="1c19afd5be4b726bb334282559cead6c35f007ededfec1ac6fb204e8ee44c586" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.973506 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c19afd5be4b726bb334282559cead6c35f007ededfec1ac6fb204e8ee44c586"} err="failed to get container status \"1c19afd5be4b726bb334282559cead6c35f007ededfec1ac6fb204e8ee44c586\": rpc error: code = NotFound desc = could not find container \"1c19afd5be4b726bb334282559cead6c35f007ededfec1ac6fb204e8ee44c586\": container with ID starting with 1c19afd5be4b726bb334282559cead6c35f007ededfec1ac6fb204e8ee44c586 not found: ID does not exist" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.973540 4799 scope.go:117] "RemoveContainer" containerID="c4cf32469a3d4feffd40673f38d4a39abb101d5196c09641054e3eff516aa22e" Nov 29 04:41:37 crc kubenswrapper[4799]: E1129 04:41:37.973804 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4cf32469a3d4feffd40673f38d4a39abb101d5196c09641054e3eff516aa22e\": container with ID starting with c4cf32469a3d4feffd40673f38d4a39abb101d5196c09641054e3eff516aa22e not found: ID does not exist" containerID="c4cf32469a3d4feffd40673f38d4a39abb101d5196c09641054e3eff516aa22e" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.973839 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4cf32469a3d4feffd40673f38d4a39abb101d5196c09641054e3eff516aa22e"} err="failed to get container status \"c4cf32469a3d4feffd40673f38d4a39abb101d5196c09641054e3eff516aa22e\": rpc error: code = NotFound desc = could not find container \"c4cf32469a3d4feffd40673f38d4a39abb101d5196c09641054e3eff516aa22e\": container with ID starting with c4cf32469a3d4feffd40673f38d4a39abb101d5196c09641054e3eff516aa22e not found: ID does not exist" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.973854 4799 scope.go:117] "RemoveContainer" containerID="03dd3900fb4cd09590cdf95e591712a29201592633fc220927a9cda4e1806f23" Nov 29 04:41:37 crc kubenswrapper[4799]: I1129 04:41:37.990180 4799 scope.go:117] "RemoveContainer" containerID="cf73277195389762deb95a92e3bc34711b22d193b6a676db03a9fc1a80c521b2" Nov 29 04:41:38 crc kubenswrapper[4799]: I1129 04:41:38.021803 4799 scope.go:117] "RemoveContainer" containerID="ffa7c3730d637df4273cd6ee511ae0064adcaee9e43dfafc39d7156a78f96706" Nov 29 04:41:38 crc kubenswrapper[4799]: I1129 04:41:38.666127 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a2cc255-bc5f-4253-bc2b-e74817971afe" path="/var/lib/kubelet/pods/8a2cc255-bc5f-4253-bc2b-e74817971afe/volumes" Nov 29 04:41:38 crc kubenswrapper[4799]: I1129 04:41:38.666751 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0bccaa4-e99f-47b1-bf20-2895f289ffd7" path="/var/lib/kubelet/pods/b0bccaa4-e99f-47b1-bf20-2895f289ffd7/volumes" Nov 29 04:41:38 crc kubenswrapper[4799]: I1129 04:41:38.667468 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d" path="/var/lib/kubelet/pods/bbfed4a1-958d-4df4-a4f8-9ce44a7d4b6d/volumes" Nov 29 04:41:38 crc kubenswrapper[4799]: I1129 04:41:38.903386 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-75c755456-6z6nn" event={"ID":"7ed02282-fade-4fad-b82e-cda665ec7998","Type":"ContainerStarted","Data":"d61fc3b4f265118b1b600ad3a1523dd939170dc2bafd24d0ea7566b45aa227f6"} Nov 29 04:41:38 crc kubenswrapper[4799]: I1129 04:41:38.903829 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-75c755456-6z6nn" Nov 29 04:41:38 crc kubenswrapper[4799]: I1129 04:41:38.909221 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-75c755456-6z6nn" Nov 29 04:41:38 crc kubenswrapper[4799]: I1129 04:41:38.927781 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-75c755456-6z6nn" podStartSLOduration=28.927752174 podStartE2EDuration="28.927752174s" podCreationTimestamp="2025-11-29 04:41:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:41:38.925654365 +0000 UTC m=+174.568584835" watchObservedRunningTime="2025-11-29 04:41:38.927752174 +0000 UTC m=+174.570682574" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.823514 4799 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 29 04:41:51 crc kubenswrapper[4799]: E1129 04:41:51.824565 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0bccaa4-e99f-47b1-bf20-2895f289ffd7" containerName="registry-server" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.824582 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0bccaa4-e99f-47b1-bf20-2895f289ffd7" containerName="registry-server" Nov 29 04:41:51 crc kubenswrapper[4799]: E1129 04:41:51.824606 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a2cc255-bc5f-4253-bc2b-e74817971afe" containerName="extract-utilities" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.824616 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a2cc255-bc5f-4253-bc2b-e74817971afe" containerName="extract-utilities" Nov 29 04:41:51 crc kubenswrapper[4799]: E1129 04:41:51.824639 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a2cc255-bc5f-4253-bc2b-e74817971afe" containerName="registry-server" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.824648 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a2cc255-bc5f-4253-bc2b-e74817971afe" containerName="registry-server" Nov 29 04:41:51 crc kubenswrapper[4799]: E1129 04:41:51.824659 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0bccaa4-e99f-47b1-bf20-2895f289ffd7" containerName="extract-content" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.824670 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0bccaa4-e99f-47b1-bf20-2895f289ffd7" containerName="extract-content" Nov 29 04:41:51 crc kubenswrapper[4799]: E1129 04:41:51.824679 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0bccaa4-e99f-47b1-bf20-2895f289ffd7" containerName="extract-utilities" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.824688 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0bccaa4-e99f-47b1-bf20-2895f289ffd7" containerName="extract-utilities" Nov 29 04:41:51 crc kubenswrapper[4799]: E1129 04:41:51.824697 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a2cc255-bc5f-4253-bc2b-e74817971afe" containerName="extract-content" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.824706 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a2cc255-bc5f-4253-bc2b-e74817971afe" containerName="extract-content" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.824859 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0bccaa4-e99f-47b1-bf20-2895f289ffd7" containerName="registry-server" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.824879 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a2cc255-bc5f-4253-bc2b-e74817971afe" containerName="registry-server" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.825342 4799 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.825530 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.825668 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea" gracePeriod=15 Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.826155 4799 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.826272 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://7079548fe9529583a1c129f57838c24cf06d1cb55a97c05c63a05d0cc2e7ec9b" gracePeriod=15 Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.826336 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc" gracePeriod=15 Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.826416 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2" gracePeriod=15 Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.826465 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe" gracePeriod=15 Nov 29 04:41:51 crc kubenswrapper[4799]: E1129 04:41:51.826569 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.826592 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Nov 29 04:41:51 crc kubenswrapper[4799]: E1129 04:41:51.826612 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.826626 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 29 04:41:51 crc kubenswrapper[4799]: E1129 04:41:51.826650 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.826665 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 29 04:41:51 crc kubenswrapper[4799]: E1129 04:41:51.826680 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.826692 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 29 04:41:51 crc kubenswrapper[4799]: E1129 04:41:51.826712 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.826726 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 29 04:41:51 crc kubenswrapper[4799]: E1129 04:41:51.826746 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.826758 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 29 04:41:51 crc kubenswrapper[4799]: E1129 04:41:51.826774 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.826813 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.827007 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.827030 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.827049 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.827066 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.827081 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.827095 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.831781 4799 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.849729 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.849839 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.849939 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.850001 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.850062 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.850128 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.850162 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.850259 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.887659 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.911168 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.912443 4799 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.913414 4799 status_manager.go:851] "Failed to get status for pod" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" pod="openshift-network-diagnostics/network-check-target-xd92c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/pods/network-check-target-xd92c\": dial tcp 38.102.83.83:6443: connect: connection refused" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.951701 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.951773 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.951834 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.951873 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.951895 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.951926 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.951957 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.951988 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.952096 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.954965 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.955020 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.955048 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.955072 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.955095 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.955119 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 29 04:41:51 crc kubenswrapper[4799]: I1129 04:41:51.955141 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 29 04:41:52 crc kubenswrapper[4799]: I1129 04:41:51.994528 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 29 04:41:52 crc kubenswrapper[4799]: I1129 04:41:51.996906 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 29 04:41:52 crc kubenswrapper[4799]: I1129 04:41:51.998163 4799 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="7079548fe9529583a1c129f57838c24cf06d1cb55a97c05c63a05d0cc2e7ec9b" exitCode=0 Nov 29 04:41:52 crc kubenswrapper[4799]: I1129 04:41:51.998209 4799 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc" exitCode=0 Nov 29 04:41:52 crc kubenswrapper[4799]: I1129 04:41:51.998237 4799 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2" exitCode=0 Nov 29 04:41:52 crc kubenswrapper[4799]: I1129 04:41:51.998246 4799 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe" exitCode=2 Nov 29 04:41:52 crc kubenswrapper[4799]: I1129 04:41:51.998306 4799 scope.go:117] "RemoveContainer" containerID="d7762237003a5a256ff20a16a4869462e2e0dd1426aa391f7ec74ed9b1f572d4" Nov 29 04:41:52 crc kubenswrapper[4799]: I1129 04:41:52.176239 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 29 04:41:52 crc kubenswrapper[4799]: W1129 04:41:52.198216 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-5b2702f2ac3e207b7f3f7c130f1995088277e01cc9d2ccaa1a6a7f003eb47267 WatchSource:0}: Error finding container 5b2702f2ac3e207b7f3f7c130f1995088277e01cc9d2ccaa1a6a7f003eb47267: Status 404 returned error can't find the container with id 5b2702f2ac3e207b7f3f7c130f1995088277e01cc9d2ccaa1a6a7f003eb47267 Nov 29 04:41:52 crc kubenswrapper[4799]: E1129 04:41:52.203530 4799 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.83:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187c608ce09cd8e4 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-29 04:41:52.202455268 +0000 UTC m=+187.845385668,LastTimestamp:2025-11-29 04:41:52.202455268 +0000 UTC m=+187.845385668,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 29 04:41:53 crc kubenswrapper[4799]: I1129 04:41:53.006159 4799 generic.go:334] "Generic (PLEG): container finished" podID="7dae4078-40b8-46a7-a83c-d162a1295c27" containerID="632a00da1e2b592990cf5d9f4cc02d6562ee52437f23783ed9baf4dfdd2120ca" exitCode=0 Nov 29 04:41:53 crc kubenswrapper[4799]: I1129 04:41:53.006286 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"7dae4078-40b8-46a7-a83c-d162a1295c27","Type":"ContainerDied","Data":"632a00da1e2b592990cf5d9f4cc02d6562ee52437f23783ed9baf4dfdd2120ca"} Nov 29 04:41:53 crc kubenswrapper[4799]: I1129 04:41:53.007773 4799 status_manager.go:851] "Failed to get status for pod" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" pod="openshift-network-diagnostics/network-check-target-xd92c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/pods/network-check-target-xd92c\": dial tcp 38.102.83.83:6443: connect: connection refused" Nov 29 04:41:53 crc kubenswrapper[4799]: I1129 04:41:53.008079 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"71b4c7ac05233cd3efa579f07ad67706ced0921ba20e8ec60cb539dcde2a0ec7"} Nov 29 04:41:53 crc kubenswrapper[4799]: I1129 04:41:53.008129 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"5b2702f2ac3e207b7f3f7c130f1995088277e01cc9d2ccaa1a6a7f003eb47267"} Nov 29 04:41:53 crc kubenswrapper[4799]: I1129 04:41:53.008282 4799 status_manager.go:851] "Failed to get status for pod" podUID="7dae4078-40b8-46a7-a83c-d162a1295c27" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Nov 29 04:41:53 crc kubenswrapper[4799]: I1129 04:41:53.008865 4799 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Nov 29 04:41:53 crc kubenswrapper[4799]: I1129 04:41:53.009559 4799 status_manager.go:851] "Failed to get status for pod" podUID="7dae4078-40b8-46a7-a83c-d162a1295c27" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Nov 29 04:41:53 crc kubenswrapper[4799]: I1129 04:41:53.010014 4799 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Nov 29 04:41:53 crc kubenswrapper[4799]: I1129 04:41:53.010858 4799 status_manager.go:851] "Failed to get status for pod" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" pod="openshift-network-diagnostics/network-check-target-xd92c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/pods/network-check-target-xd92c\": dial tcp 38.102.83.83:6443: connect: connection refused" Nov 29 04:41:53 crc kubenswrapper[4799]: I1129 04:41:53.011781 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 29 04:41:54 crc kubenswrapper[4799]: I1129 04:41:54.295929 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 29 04:41:54 crc kubenswrapper[4799]: I1129 04:41:54.297347 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 04:41:54 crc kubenswrapper[4799]: I1129 04:41:54.298012 4799 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Nov 29 04:41:54 crc kubenswrapper[4799]: I1129 04:41:54.298326 4799 status_manager.go:851] "Failed to get status for pod" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" pod="openshift-network-diagnostics/network-check-target-xd92c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/pods/network-check-target-xd92c\": dial tcp 38.102.83.83:6443: connect: connection refused" Nov 29 04:41:54 crc kubenswrapper[4799]: I1129 04:41:54.298844 4799 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Nov 29 04:41:54 crc kubenswrapper[4799]: I1129 04:41:54.299131 4799 status_manager.go:851] "Failed to get status for pod" podUID="7dae4078-40b8-46a7-a83c-d162a1295c27" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Nov 29 04:41:54 crc kubenswrapper[4799]: I1129 04:41:54.300552 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 29 04:41:54 crc kubenswrapper[4799]: I1129 04:41:54.301208 4799 status_manager.go:851] "Failed to get status for pod" podUID="7dae4078-40b8-46a7-a83c-d162a1295c27" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Nov 29 04:41:54 crc kubenswrapper[4799]: I1129 04:41:54.302078 4799 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Nov 29 04:41:54 crc kubenswrapper[4799]: I1129 04:41:54.302428 4799 status_manager.go:851] "Failed to get status for pod" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" pod="openshift-network-diagnostics/network-check-target-xd92c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/pods/network-check-target-xd92c\": dial tcp 38.102.83.83:6443: connect: connection refused" Nov 29 04:41:54 crc kubenswrapper[4799]: I1129 04:41:54.302820 4799 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Nov 29 04:41:54 crc kubenswrapper[4799]: I1129 04:41:54.431301 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7dae4078-40b8-46a7-a83c-d162a1295c27-kube-api-access\") pod \"7dae4078-40b8-46a7-a83c-d162a1295c27\" (UID: \"7dae4078-40b8-46a7-a83c-d162a1295c27\") " Nov 29 04:41:54 crc kubenswrapper[4799]: I1129 04:41:54.432222 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 29 04:41:54 crc kubenswrapper[4799]: I1129 04:41:54.432255 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 29 04:41:54 crc kubenswrapper[4799]: I1129 04:41:54.432289 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7dae4078-40b8-46a7-a83c-d162a1295c27-kubelet-dir\") pod \"7dae4078-40b8-46a7-a83c-d162a1295c27\" (UID: \"7dae4078-40b8-46a7-a83c-d162a1295c27\") " Nov 29 04:41:54 crc kubenswrapper[4799]: I1129 04:41:54.432393 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/7dae4078-40b8-46a7-a83c-d162a1295c27-var-lock\") pod \"7dae4078-40b8-46a7-a83c-d162a1295c27\" (UID: \"7dae4078-40b8-46a7-a83c-d162a1295c27\") " Nov 29 04:41:54 crc kubenswrapper[4799]: I1129 04:41:54.432416 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 29 04:41:54 crc kubenswrapper[4799]: I1129 04:41:54.432718 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 04:41:54 crc kubenswrapper[4799]: I1129 04:41:54.432754 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7dae4078-40b8-46a7-a83c-d162a1295c27-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "7dae4078-40b8-46a7-a83c-d162a1295c27" (UID: "7dae4078-40b8-46a7-a83c-d162a1295c27"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 04:41:54 crc kubenswrapper[4799]: I1129 04:41:54.432782 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7dae4078-40b8-46a7-a83c-d162a1295c27-var-lock" (OuterVolumeSpecName: "var-lock") pod "7dae4078-40b8-46a7-a83c-d162a1295c27" (UID: "7dae4078-40b8-46a7-a83c-d162a1295c27"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 04:41:54 crc kubenswrapper[4799]: I1129 04:41:54.432919 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 04:41:54 crc kubenswrapper[4799]: I1129 04:41:54.433005 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 04:41:54 crc kubenswrapper[4799]: I1129 04:41:54.438940 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7dae4078-40b8-46a7-a83c-d162a1295c27-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "7dae4078-40b8-46a7-a83c-d162a1295c27" (UID: "7dae4078-40b8-46a7-a83c-d162a1295c27"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:41:54 crc kubenswrapper[4799]: I1129 04:41:54.533617 4799 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 29 04:41:54 crc kubenswrapper[4799]: I1129 04:41:54.533685 4799 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 29 04:41:54 crc kubenswrapper[4799]: I1129 04:41:54.533696 4799 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7dae4078-40b8-46a7-a83c-d162a1295c27-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 29 04:41:54 crc kubenswrapper[4799]: I1129 04:41:54.533707 4799 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/7dae4078-40b8-46a7-a83c-d162a1295c27-var-lock\") on node \"crc\" DevicePath \"\"" Nov 29 04:41:54 crc kubenswrapper[4799]: I1129 04:41:54.533715 4799 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Nov 29 04:41:54 crc kubenswrapper[4799]: I1129 04:41:54.533724 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7dae4078-40b8-46a7-a83c-d162a1295c27-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 29 04:41:54 crc kubenswrapper[4799]: I1129 04:41:54.664395 4799 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Nov 29 04:41:54 crc kubenswrapper[4799]: I1129 04:41:54.665009 4799 status_manager.go:851] "Failed to get status for pod" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" pod="openshift-network-diagnostics/network-check-target-xd92c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/pods/network-check-target-xd92c\": dial tcp 38.102.83.83:6443: connect: connection refused" Nov 29 04:41:54 crc kubenswrapper[4799]: I1129 04:41:54.665286 4799 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Nov 29 04:41:54 crc kubenswrapper[4799]: I1129 04:41:54.665644 4799 status_manager.go:851] "Failed to get status for pod" podUID="7dae4078-40b8-46a7-a83c-d162a1295c27" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Nov 29 04:41:54 crc kubenswrapper[4799]: I1129 04:41:54.666883 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Nov 29 04:41:55 crc kubenswrapper[4799]: I1129 04:41:55.027355 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 29 04:41:55 crc kubenswrapper[4799]: I1129 04:41:55.027375 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"7dae4078-40b8-46a7-a83c-d162a1295c27","Type":"ContainerDied","Data":"04d95f44ec2c81dd01c02f291e6a1d455016e35effdf1c9a6fb92558e72c4838"} Nov 29 04:41:55 crc kubenswrapper[4799]: I1129 04:41:55.027433 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="04d95f44ec2c81dd01c02f291e6a1d455016e35effdf1c9a6fb92558e72c4838" Nov 29 04:41:55 crc kubenswrapper[4799]: I1129 04:41:55.031210 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 29 04:41:55 crc kubenswrapper[4799]: I1129 04:41:55.032007 4799 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea" exitCode=0 Nov 29 04:41:55 crc kubenswrapper[4799]: I1129 04:41:55.032132 4799 scope.go:117] "RemoveContainer" containerID="7079548fe9529583a1c129f57838c24cf06d1cb55a97c05c63a05d0cc2e7ec9b" Nov 29 04:41:55 crc kubenswrapper[4799]: I1129 04:41:55.032253 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 04:41:55 crc kubenswrapper[4799]: I1129 04:41:55.033141 4799 status_manager.go:851] "Failed to get status for pod" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" pod="openshift-network-diagnostics/network-check-target-xd92c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/pods/network-check-target-xd92c\": dial tcp 38.102.83.83:6443: connect: connection refused" Nov 29 04:41:55 crc kubenswrapper[4799]: I1129 04:41:55.033872 4799 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Nov 29 04:41:55 crc kubenswrapper[4799]: I1129 04:41:55.034334 4799 status_manager.go:851] "Failed to get status for pod" podUID="7dae4078-40b8-46a7-a83c-d162a1295c27" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Nov 29 04:41:55 crc kubenswrapper[4799]: I1129 04:41:55.034771 4799 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Nov 29 04:41:55 crc kubenswrapper[4799]: I1129 04:41:55.035461 4799 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Nov 29 04:41:55 crc kubenswrapper[4799]: I1129 04:41:55.036190 4799 status_manager.go:851] "Failed to get status for pod" podUID="7dae4078-40b8-46a7-a83c-d162a1295c27" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Nov 29 04:41:55 crc kubenswrapper[4799]: I1129 04:41:55.037000 4799 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Nov 29 04:41:55 crc kubenswrapper[4799]: I1129 04:41:55.037446 4799 status_manager.go:851] "Failed to get status for pod" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" pod="openshift-network-diagnostics/network-check-target-xd92c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/pods/network-check-target-xd92c\": dial tcp 38.102.83.83:6443: connect: connection refused" Nov 29 04:41:55 crc kubenswrapper[4799]: I1129 04:41:55.038000 4799 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Nov 29 04:41:55 crc kubenswrapper[4799]: I1129 04:41:55.038293 4799 status_manager.go:851] "Failed to get status for pod" podUID="7dae4078-40b8-46a7-a83c-d162a1295c27" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Nov 29 04:41:55 crc kubenswrapper[4799]: I1129 04:41:55.038613 4799 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Nov 29 04:41:55 crc kubenswrapper[4799]: I1129 04:41:55.039345 4799 status_manager.go:851] "Failed to get status for pod" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" pod="openshift-network-diagnostics/network-check-target-xd92c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/pods/network-check-target-xd92c\": dial tcp 38.102.83.83:6443: connect: connection refused" Nov 29 04:41:55 crc kubenswrapper[4799]: I1129 04:41:55.053755 4799 scope.go:117] "RemoveContainer" containerID="2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc" Nov 29 04:41:55 crc kubenswrapper[4799]: I1129 04:41:55.077500 4799 scope.go:117] "RemoveContainer" containerID="bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2" Nov 29 04:41:55 crc kubenswrapper[4799]: I1129 04:41:55.094243 4799 scope.go:117] "RemoveContainer" containerID="fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe" Nov 29 04:41:55 crc kubenswrapper[4799]: I1129 04:41:55.118029 4799 scope.go:117] "RemoveContainer" containerID="0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea" Nov 29 04:41:55 crc kubenswrapper[4799]: I1129 04:41:55.142305 4799 scope.go:117] "RemoveContainer" containerID="d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f" Nov 29 04:41:55 crc kubenswrapper[4799]: I1129 04:41:55.163837 4799 scope.go:117] "RemoveContainer" containerID="7079548fe9529583a1c129f57838c24cf06d1cb55a97c05c63a05d0cc2e7ec9b" Nov 29 04:41:55 crc kubenswrapper[4799]: E1129 04:41:55.166532 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7079548fe9529583a1c129f57838c24cf06d1cb55a97c05c63a05d0cc2e7ec9b\": container with ID starting with 7079548fe9529583a1c129f57838c24cf06d1cb55a97c05c63a05d0cc2e7ec9b not found: ID does not exist" containerID="7079548fe9529583a1c129f57838c24cf06d1cb55a97c05c63a05d0cc2e7ec9b" Nov 29 04:41:55 crc kubenswrapper[4799]: I1129 04:41:55.166602 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7079548fe9529583a1c129f57838c24cf06d1cb55a97c05c63a05d0cc2e7ec9b"} err="failed to get container status \"7079548fe9529583a1c129f57838c24cf06d1cb55a97c05c63a05d0cc2e7ec9b\": rpc error: code = NotFound desc = could not find container \"7079548fe9529583a1c129f57838c24cf06d1cb55a97c05c63a05d0cc2e7ec9b\": container with ID starting with 7079548fe9529583a1c129f57838c24cf06d1cb55a97c05c63a05d0cc2e7ec9b not found: ID does not exist" Nov 29 04:41:55 crc kubenswrapper[4799]: I1129 04:41:55.166648 4799 scope.go:117] "RemoveContainer" containerID="2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc" Nov 29 04:41:55 crc kubenswrapper[4799]: E1129 04:41:55.168019 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc\": container with ID starting with 2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc not found: ID does not exist" containerID="2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc" Nov 29 04:41:55 crc kubenswrapper[4799]: I1129 04:41:55.168114 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc"} err="failed to get container status \"2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc\": rpc error: code = NotFound desc = could not find container \"2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc\": container with ID starting with 2a7d5f837dea66e49525c01663ce5a46b68e8c71c5b412f2f531f17c3e1978fc not found: ID does not exist" Nov 29 04:41:55 crc kubenswrapper[4799]: I1129 04:41:55.168156 4799 scope.go:117] "RemoveContainer" containerID="bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2" Nov 29 04:41:55 crc kubenswrapper[4799]: E1129 04:41:55.168842 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2\": container with ID starting with bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2 not found: ID does not exist" containerID="bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2" Nov 29 04:41:55 crc kubenswrapper[4799]: I1129 04:41:55.168927 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2"} err="failed to get container status \"bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2\": rpc error: code = NotFound desc = could not find container \"bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2\": container with ID starting with bbbbb791b0e084370f08c78a9290c383a880785193c4814293d7bfa655e7eed2 not found: ID does not exist" Nov 29 04:41:55 crc kubenswrapper[4799]: I1129 04:41:55.168967 4799 scope.go:117] "RemoveContainer" containerID="fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe" Nov 29 04:41:55 crc kubenswrapper[4799]: E1129 04:41:55.169985 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe\": container with ID starting with fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe not found: ID does not exist" containerID="fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe" Nov 29 04:41:55 crc kubenswrapper[4799]: I1129 04:41:55.170115 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe"} err="failed to get container status \"fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe\": rpc error: code = NotFound desc = could not find container \"fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe\": container with ID starting with fbffaaf28469e6abd639ad1dfbb352bbec0d02befeae503598ab9769685d8cfe not found: ID does not exist" Nov 29 04:41:55 crc kubenswrapper[4799]: I1129 04:41:55.170194 4799 scope.go:117] "RemoveContainer" containerID="0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea" Nov 29 04:41:55 crc kubenswrapper[4799]: E1129 04:41:55.171941 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea\": container with ID starting with 0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea not found: ID does not exist" containerID="0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea" Nov 29 04:41:55 crc kubenswrapper[4799]: I1129 04:41:55.171992 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea"} err="failed to get container status \"0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea\": rpc error: code = NotFound desc = could not find container \"0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea\": container with ID starting with 0ee1de9a7e171f179f84d6019f52dc7f5265635b2c2e278098314ba2d2f208ea not found: ID does not exist" Nov 29 04:41:55 crc kubenswrapper[4799]: I1129 04:41:55.172028 4799 scope.go:117] "RemoveContainer" containerID="d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f" Nov 29 04:41:55 crc kubenswrapper[4799]: E1129 04:41:55.172505 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\": container with ID starting with d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f not found: ID does not exist" containerID="d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f" Nov 29 04:41:55 crc kubenswrapper[4799]: I1129 04:41:55.172636 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f"} err="failed to get container status \"d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\": rpc error: code = NotFound desc = could not find container \"d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f\": container with ID starting with d2a13cbd1dca5c9cf8cdaec89b18500578b82c2bf61372f90c88710e68c30d2f not found: ID does not exist" Nov 29 04:41:56 crc kubenswrapper[4799]: E1129 04:41:56.568462 4799 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.83:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187c608ce09cd8e4 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-29 04:41:52.202455268 +0000 UTC m=+187.845385668,LastTimestamp:2025-11-29 04:41:52.202455268 +0000 UTC m=+187.845385668,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 29 04:41:56 crc kubenswrapper[4799]: E1129 04:41:56.774411 4799 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.83:6443: connect: connection refused" Nov 29 04:41:56 crc kubenswrapper[4799]: E1129 04:41:56.775194 4799 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.83:6443: connect: connection refused" Nov 29 04:41:56 crc kubenswrapper[4799]: E1129 04:41:56.775658 4799 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.83:6443: connect: connection refused" Nov 29 04:41:56 crc kubenswrapper[4799]: E1129 04:41:56.776045 4799 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.83:6443: connect: connection refused" Nov 29 04:41:56 crc kubenswrapper[4799]: E1129 04:41:56.776324 4799 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.83:6443: connect: connection refused" Nov 29 04:41:56 crc kubenswrapper[4799]: I1129 04:41:56.776377 4799 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Nov 29 04:41:56 crc kubenswrapper[4799]: E1129 04:41:56.776618 4799 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.83:6443: connect: connection refused" interval="200ms" Nov 29 04:41:56 crc kubenswrapper[4799]: E1129 04:41:56.978081 4799 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.83:6443: connect: connection refused" interval="400ms" Nov 29 04:41:57 crc kubenswrapper[4799]: E1129 04:41:57.378882 4799 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.83:6443: connect: connection refused" interval="800ms" Nov 29 04:41:58 crc kubenswrapper[4799]: E1129 04:41:58.180828 4799 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.83:6443: connect: connection refused" interval="1.6s" Nov 29 04:41:59 crc kubenswrapper[4799]: E1129 04:41:59.782222 4799 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.83:6443: connect: connection refused" interval="3.2s" Nov 29 04:42:02 crc kubenswrapper[4799]: E1129 04:42:02.984591 4799 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.83:6443: connect: connection refused" interval="6.4s" Nov 29 04:42:04 crc kubenswrapper[4799]: I1129 04:42:04.663305 4799 status_manager.go:851] "Failed to get status for pod" podUID="7dae4078-40b8-46a7-a83c-d162a1295c27" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Nov 29 04:42:04 crc kubenswrapper[4799]: I1129 04:42:04.664864 4799 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Nov 29 04:42:04 crc kubenswrapper[4799]: I1129 04:42:04.665256 4799 status_manager.go:851] "Failed to get status for pod" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" pod="openshift-network-diagnostics/network-check-target-xd92c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/pods/network-check-target-xd92c\": dial tcp 38.102.83.83:6443: connect: connection refused" Nov 29 04:42:06 crc kubenswrapper[4799]: E1129 04:42:06.573499 4799 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.83:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187c608ce09cd8e4 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-29 04:41:52.202455268 +0000 UTC m=+187.845385668,LastTimestamp:2025-11-29 04:41:52.202455268 +0000 UTC m=+187.845385668,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 29 04:42:06 crc kubenswrapper[4799]: I1129 04:42:06.658776 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 04:42:06 crc kubenswrapper[4799]: I1129 04:42:06.660916 4799 status_manager.go:851] "Failed to get status for pod" podUID="7dae4078-40b8-46a7-a83c-d162a1295c27" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Nov 29 04:42:06 crc kubenswrapper[4799]: I1129 04:42:06.661988 4799 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Nov 29 04:42:06 crc kubenswrapper[4799]: I1129 04:42:06.662958 4799 status_manager.go:851] "Failed to get status for pod" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" pod="openshift-network-diagnostics/network-check-target-xd92c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/pods/network-check-target-xd92c\": dial tcp 38.102.83.83:6443: connect: connection refused" Nov 29 04:42:06 crc kubenswrapper[4799]: I1129 04:42:06.678014 4799 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4d40160e-075f-44b7-9d9d-3291e7774f27" Nov 29 04:42:06 crc kubenswrapper[4799]: I1129 04:42:06.678070 4799 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4d40160e-075f-44b7-9d9d-3291e7774f27" Nov 29 04:42:06 crc kubenswrapper[4799]: E1129 04:42:06.678870 4799 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 04:42:06 crc kubenswrapper[4799]: I1129 04:42:06.679647 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 04:42:07 crc kubenswrapper[4799]: I1129 04:42:07.122067 4799 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="6849009585e3dee1de9d2f7c8cc495f29b4df036da97a952e8ff38aed89ed663" exitCode=0 Nov 29 04:42:07 crc kubenswrapper[4799]: I1129 04:42:07.122222 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"6849009585e3dee1de9d2f7c8cc495f29b4df036da97a952e8ff38aed89ed663"} Nov 29 04:42:07 crc kubenswrapper[4799]: I1129 04:42:07.122520 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"64246569c849ece4ed8c3459d6c12892b72fea9623cf89399789f87f7a96a52b"} Nov 29 04:42:07 crc kubenswrapper[4799]: I1129 04:42:07.122840 4799 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4d40160e-075f-44b7-9d9d-3291e7774f27" Nov 29 04:42:07 crc kubenswrapper[4799]: I1129 04:42:07.122855 4799 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4d40160e-075f-44b7-9d9d-3291e7774f27" Nov 29 04:42:07 crc kubenswrapper[4799]: I1129 04:42:07.123486 4799 status_manager.go:851] "Failed to get status for pod" podUID="7dae4078-40b8-46a7-a83c-d162a1295c27" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Nov 29 04:42:07 crc kubenswrapper[4799]: E1129 04:42:07.123966 4799 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 04:42:07 crc kubenswrapper[4799]: I1129 04:42:07.124003 4799 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Nov 29 04:42:07 crc kubenswrapper[4799]: I1129 04:42:07.124422 4799 status_manager.go:851] "Failed to get status for pod" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" pod="openshift-network-diagnostics/network-check-target-xd92c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/pods/network-check-target-xd92c\": dial tcp 38.102.83.83:6443: connect: connection refused" Nov 29 04:42:07 crc kubenswrapper[4799]: I1129 04:42:07.126722 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 29 04:42:07 crc kubenswrapper[4799]: I1129 04:42:07.126823 4799 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="ea4c241d29297a73dce12bd0bc2863c56475eb35b8fe8c93991ba0ef8695e38f" exitCode=1 Nov 29 04:42:07 crc kubenswrapper[4799]: I1129 04:42:07.126877 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"ea4c241d29297a73dce12bd0bc2863c56475eb35b8fe8c93991ba0ef8695e38f"} Nov 29 04:42:07 crc kubenswrapper[4799]: I1129 04:42:07.127544 4799 scope.go:117] "RemoveContainer" containerID="ea4c241d29297a73dce12bd0bc2863c56475eb35b8fe8c93991ba0ef8695e38f" Nov 29 04:42:07 crc kubenswrapper[4799]: I1129 04:42:07.127605 4799 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Nov 29 04:42:07 crc kubenswrapper[4799]: I1129 04:42:07.127812 4799 status_manager.go:851] "Failed to get status for pod" podUID="7dae4078-40b8-46a7-a83c-d162a1295c27" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Nov 29 04:42:07 crc kubenswrapper[4799]: I1129 04:42:07.128087 4799 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Nov 29 04:42:07 crc kubenswrapper[4799]: I1129 04:42:07.128682 4799 status_manager.go:851] "Failed to get status for pod" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" pod="openshift-network-diagnostics/network-check-target-xd92c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/pods/network-check-target-xd92c\": dial tcp 38.102.83.83:6443: connect: connection refused" Nov 29 04:42:07 crc kubenswrapper[4799]: I1129 04:42:07.498047 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 04:42:07 crc kubenswrapper[4799]: I1129 04:42:07.498110 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 04:42:07 crc kubenswrapper[4799]: I1129 04:42:07.660515 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 29 04:42:08 crc kubenswrapper[4799]: I1129 04:42:08.175514 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"a13dc9ad73c4f4a2e681f5c14b5b1cc9b4e49a5a2c05df7df76ec20b77836651"} Nov 29 04:42:08 crc kubenswrapper[4799]: I1129 04:42:08.176181 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"39295732ecf59a7833d82733853fd813498602b2f7206dd3c980187a3841b397"} Nov 29 04:42:08 crc kubenswrapper[4799]: I1129 04:42:08.176198 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"387d58fb4de276b430e0085d81737414d26e2d48d5b7cf9b340b517f9e7bfc93"} Nov 29 04:42:08 crc kubenswrapper[4799]: I1129 04:42:08.180482 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 29 04:42:08 crc kubenswrapper[4799]: I1129 04:42:08.180542 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"3fed141386a75c80b0b5ec2def5239bfae108d51a6058501695184c3ca3006f3"} Nov 29 04:42:08 crc kubenswrapper[4799]: I1129 04:42:08.265131 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 29 04:42:09 crc kubenswrapper[4799]: I1129 04:42:09.190197 4799 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4d40160e-075f-44b7-9d9d-3291e7774f27" Nov 29 04:42:09 crc kubenswrapper[4799]: I1129 04:42:09.190231 4799 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4d40160e-075f-44b7-9d9d-3291e7774f27" Nov 29 04:42:09 crc kubenswrapper[4799]: I1129 04:42:09.190455 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"e7b8fd9cf44d8137b1347761cc592f5853c7b0fd16f167d19f0266247da6ac79"} Nov 29 04:42:09 crc kubenswrapper[4799]: I1129 04:42:09.190484 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"638bb84f8e5c10cae96f77ddbbce8c11823e07a3fbd1e377a5f3952e00709dc2"} Nov 29 04:42:09 crc kubenswrapper[4799]: I1129 04:42:09.190517 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 04:42:11 crc kubenswrapper[4799]: I1129 04:42:11.680663 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 04:42:11 crc kubenswrapper[4799]: I1129 04:42:11.681188 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 04:42:11 crc kubenswrapper[4799]: I1129 04:42:11.688969 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 04:42:14 crc kubenswrapper[4799]: I1129 04:42:14.203250 4799 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 04:42:14 crc kubenswrapper[4799]: I1129 04:42:14.247601 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 29 04:42:14 crc kubenswrapper[4799]: I1129 04:42:14.254311 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 29 04:42:14 crc kubenswrapper[4799]: I1129 04:42:14.697338 4799 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="a54396ac-bf19-469c-bf65-c4b7c2d50c68" Nov 29 04:42:15 crc kubenswrapper[4799]: I1129 04:42:15.226012 4799 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4d40160e-075f-44b7-9d9d-3291e7774f27" Nov 29 04:42:15 crc kubenswrapper[4799]: I1129 04:42:15.226060 4799 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4d40160e-075f-44b7-9d9d-3291e7774f27" Nov 29 04:42:15 crc kubenswrapper[4799]: I1129 04:42:15.229110 4799 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="a54396ac-bf19-469c-bf65-c4b7c2d50c68" Nov 29 04:42:18 crc kubenswrapper[4799]: I1129 04:42:18.273407 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 29 04:42:23 crc kubenswrapper[4799]: I1129 04:42:23.957761 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 29 04:42:24 crc kubenswrapper[4799]: I1129 04:42:24.139229 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 29 04:42:25 crc kubenswrapper[4799]: I1129 04:42:25.033912 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 29 04:42:25 crc kubenswrapper[4799]: I1129 04:42:25.155406 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 29 04:42:25 crc kubenswrapper[4799]: I1129 04:42:25.361138 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 29 04:42:25 crc kubenswrapper[4799]: I1129 04:42:25.384426 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 29 04:42:25 crc kubenswrapper[4799]: I1129 04:42:25.415020 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 29 04:42:25 crc kubenswrapper[4799]: I1129 04:42:25.713822 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 29 04:42:25 crc kubenswrapper[4799]: I1129 04:42:25.797415 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 29 04:42:25 crc kubenswrapper[4799]: I1129 04:42:25.800376 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 29 04:42:25 crc kubenswrapper[4799]: I1129 04:42:25.868528 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 29 04:42:25 crc kubenswrapper[4799]: I1129 04:42:25.875627 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 29 04:42:26 crc kubenswrapper[4799]: I1129 04:42:26.017587 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 29 04:42:26 crc kubenswrapper[4799]: I1129 04:42:26.084990 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 29 04:42:26 crc kubenswrapper[4799]: I1129 04:42:26.168657 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 29 04:42:26 crc kubenswrapper[4799]: I1129 04:42:26.219840 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 29 04:42:26 crc kubenswrapper[4799]: I1129 04:42:26.354911 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 29 04:42:26 crc kubenswrapper[4799]: I1129 04:42:26.443823 4799 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 29 04:42:26 crc kubenswrapper[4799]: I1129 04:42:26.587687 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 29 04:42:26 crc kubenswrapper[4799]: I1129 04:42:26.686339 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 29 04:42:26 crc kubenswrapper[4799]: I1129 04:42:26.722550 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 29 04:42:26 crc kubenswrapper[4799]: I1129 04:42:26.753509 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 29 04:42:26 crc kubenswrapper[4799]: I1129 04:42:26.906888 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 29 04:42:27 crc kubenswrapper[4799]: I1129 04:42:27.134193 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 29 04:42:27 crc kubenswrapper[4799]: I1129 04:42:27.322507 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 29 04:42:27 crc kubenswrapper[4799]: I1129 04:42:27.501754 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 29 04:42:27 crc kubenswrapper[4799]: I1129 04:42:27.528722 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 29 04:42:27 crc kubenswrapper[4799]: I1129 04:42:27.601528 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 29 04:42:27 crc kubenswrapper[4799]: I1129 04:42:27.617286 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 29 04:42:27 crc kubenswrapper[4799]: I1129 04:42:27.701776 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 29 04:42:27 crc kubenswrapper[4799]: I1129 04:42:27.705449 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 29 04:42:27 crc kubenswrapper[4799]: I1129 04:42:27.707910 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 29 04:42:27 crc kubenswrapper[4799]: I1129 04:42:27.716526 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 29 04:42:27 crc kubenswrapper[4799]: I1129 04:42:27.838541 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 29 04:42:27 crc kubenswrapper[4799]: I1129 04:42:27.920123 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 29 04:42:27 crc kubenswrapper[4799]: I1129 04:42:27.936131 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 29 04:42:27 crc kubenswrapper[4799]: I1129 04:42:27.941069 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 29 04:42:28 crc kubenswrapper[4799]: I1129 04:42:28.054891 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 29 04:42:28 crc kubenswrapper[4799]: I1129 04:42:28.087357 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 29 04:42:28 crc kubenswrapper[4799]: I1129 04:42:28.135618 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 29 04:42:28 crc kubenswrapper[4799]: I1129 04:42:28.155910 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 29 04:42:28 crc kubenswrapper[4799]: I1129 04:42:28.372017 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 29 04:42:28 crc kubenswrapper[4799]: I1129 04:42:28.512570 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 29 04:42:28 crc kubenswrapper[4799]: I1129 04:42:28.618620 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 29 04:42:28 crc kubenswrapper[4799]: I1129 04:42:28.652296 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 29 04:42:28 crc kubenswrapper[4799]: I1129 04:42:28.670607 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 29 04:42:28 crc kubenswrapper[4799]: I1129 04:42:28.683431 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 29 04:42:28 crc kubenswrapper[4799]: I1129 04:42:28.740976 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 29 04:42:28 crc kubenswrapper[4799]: I1129 04:42:28.904235 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 29 04:42:29 crc kubenswrapper[4799]: I1129 04:42:29.099569 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 29 04:42:29 crc kubenswrapper[4799]: I1129 04:42:29.104960 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 29 04:42:29 crc kubenswrapper[4799]: I1129 04:42:29.112171 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 29 04:42:29 crc kubenswrapper[4799]: I1129 04:42:29.323042 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 29 04:42:29 crc kubenswrapper[4799]: I1129 04:42:29.398952 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 29 04:42:29 crc kubenswrapper[4799]: I1129 04:42:29.530868 4799 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 29 04:42:29 crc kubenswrapper[4799]: I1129 04:42:29.579655 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 29 04:42:29 crc kubenswrapper[4799]: I1129 04:42:29.976552 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 29 04:42:29 crc kubenswrapper[4799]: I1129 04:42:29.981103 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 29 04:42:29 crc kubenswrapper[4799]: I1129 04:42:29.988289 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 29 04:42:30 crc kubenswrapper[4799]: I1129 04:42:30.005601 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 29 04:42:30 crc kubenswrapper[4799]: I1129 04:42:30.043209 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 29 04:42:30 crc kubenswrapper[4799]: I1129 04:42:30.073203 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 29 04:42:30 crc kubenswrapper[4799]: I1129 04:42:30.116575 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 29 04:42:30 crc kubenswrapper[4799]: I1129 04:42:30.234088 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 29 04:42:30 crc kubenswrapper[4799]: I1129 04:42:30.272348 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 29 04:42:30 crc kubenswrapper[4799]: I1129 04:42:30.332376 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 29 04:42:30 crc kubenswrapper[4799]: I1129 04:42:30.433254 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 29 04:42:30 crc kubenswrapper[4799]: I1129 04:42:30.469067 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 29 04:42:30 crc kubenswrapper[4799]: I1129 04:42:30.516331 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 29 04:42:30 crc kubenswrapper[4799]: I1129 04:42:30.573258 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 29 04:42:30 crc kubenswrapper[4799]: I1129 04:42:30.581356 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 29 04:42:30 crc kubenswrapper[4799]: I1129 04:42:30.694518 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 29 04:42:30 crc kubenswrapper[4799]: I1129 04:42:30.737915 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 29 04:42:30 crc kubenswrapper[4799]: I1129 04:42:30.774072 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 29 04:42:31 crc kubenswrapper[4799]: I1129 04:42:31.046439 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 29 04:42:31 crc kubenswrapper[4799]: I1129 04:42:31.073383 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 29 04:42:31 crc kubenswrapper[4799]: I1129 04:42:31.123634 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 29 04:42:31 crc kubenswrapper[4799]: I1129 04:42:31.146273 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 29 04:42:31 crc kubenswrapper[4799]: I1129 04:42:31.148659 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 29 04:42:31 crc kubenswrapper[4799]: I1129 04:42:31.236872 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 29 04:42:31 crc kubenswrapper[4799]: I1129 04:42:31.255289 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 29 04:42:31 crc kubenswrapper[4799]: I1129 04:42:31.259700 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 29 04:42:31 crc kubenswrapper[4799]: I1129 04:42:31.287141 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 29 04:42:31 crc kubenswrapper[4799]: I1129 04:42:31.339590 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 29 04:42:31 crc kubenswrapper[4799]: I1129 04:42:31.342919 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 29 04:42:31 crc kubenswrapper[4799]: I1129 04:42:31.459442 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 29 04:42:31 crc kubenswrapper[4799]: I1129 04:42:31.525164 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 29 04:42:31 crc kubenswrapper[4799]: I1129 04:42:31.585895 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 29 04:42:31 crc kubenswrapper[4799]: I1129 04:42:31.588404 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 29 04:42:31 crc kubenswrapper[4799]: I1129 04:42:31.657845 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 29 04:42:31 crc kubenswrapper[4799]: I1129 04:42:31.665553 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 29 04:42:31 crc kubenswrapper[4799]: I1129 04:42:31.683227 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 29 04:42:31 crc kubenswrapper[4799]: I1129 04:42:31.738247 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 29 04:42:31 crc kubenswrapper[4799]: I1129 04:42:31.763100 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 29 04:42:31 crc kubenswrapper[4799]: I1129 04:42:31.835508 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 29 04:42:31 crc kubenswrapper[4799]: I1129 04:42:31.859428 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 29 04:42:31 crc kubenswrapper[4799]: I1129 04:42:31.919395 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 29 04:42:31 crc kubenswrapper[4799]: I1129 04:42:31.953637 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 29 04:42:31 crc kubenswrapper[4799]: I1129 04:42:31.962952 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 29 04:42:31 crc kubenswrapper[4799]: I1129 04:42:31.969982 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 29 04:42:31 crc kubenswrapper[4799]: I1129 04:42:31.974944 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 29 04:42:32 crc kubenswrapper[4799]: I1129 04:42:32.001526 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 29 04:42:32 crc kubenswrapper[4799]: I1129 04:42:32.030918 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 29 04:42:32 crc kubenswrapper[4799]: I1129 04:42:32.052087 4799 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 29 04:42:32 crc kubenswrapper[4799]: I1129 04:42:32.067014 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 29 04:42:32 crc kubenswrapper[4799]: I1129 04:42:32.091352 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 29 04:42:32 crc kubenswrapper[4799]: I1129 04:42:32.126296 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 29 04:42:32 crc kubenswrapper[4799]: I1129 04:42:32.178496 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 29 04:42:32 crc kubenswrapper[4799]: I1129 04:42:32.187347 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 29 04:42:32 crc kubenswrapper[4799]: I1129 04:42:32.264928 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 29 04:42:32 crc kubenswrapper[4799]: I1129 04:42:32.343050 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 29 04:42:32 crc kubenswrapper[4799]: I1129 04:42:32.411699 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 29 04:42:32 crc kubenswrapper[4799]: I1129 04:42:32.438419 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 29 04:42:32 crc kubenswrapper[4799]: I1129 04:42:32.440910 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 29 04:42:32 crc kubenswrapper[4799]: I1129 04:42:32.496407 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 29 04:42:32 crc kubenswrapper[4799]: I1129 04:42:32.504238 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 29 04:42:32 crc kubenswrapper[4799]: I1129 04:42:32.507068 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 29 04:42:32 crc kubenswrapper[4799]: I1129 04:42:32.533369 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 29 04:42:32 crc kubenswrapper[4799]: I1129 04:42:32.599355 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 29 04:42:32 crc kubenswrapper[4799]: I1129 04:42:32.599355 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 29 04:42:32 crc kubenswrapper[4799]: I1129 04:42:32.655270 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 29 04:42:32 crc kubenswrapper[4799]: I1129 04:42:32.671052 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 29 04:42:32 crc kubenswrapper[4799]: I1129 04:42:32.697125 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 29 04:42:32 crc kubenswrapper[4799]: I1129 04:42:32.704236 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 29 04:42:32 crc kubenswrapper[4799]: I1129 04:42:32.778772 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 29 04:42:32 crc kubenswrapper[4799]: I1129 04:42:32.888895 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 29 04:42:32 crc kubenswrapper[4799]: I1129 04:42:32.906648 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 29 04:42:32 crc kubenswrapper[4799]: I1129 04:42:32.971585 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 29 04:42:32 crc kubenswrapper[4799]: I1129 04:42:32.982684 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 29 04:42:33 crc kubenswrapper[4799]: I1129 04:42:33.109528 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 29 04:42:33 crc kubenswrapper[4799]: I1129 04:42:33.145049 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 29 04:42:33 crc kubenswrapper[4799]: I1129 04:42:33.215902 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 29 04:42:33 crc kubenswrapper[4799]: I1129 04:42:33.222956 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 29 04:42:33 crc kubenswrapper[4799]: I1129 04:42:33.295420 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 29 04:42:33 crc kubenswrapper[4799]: I1129 04:42:33.310391 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 29 04:42:33 crc kubenswrapper[4799]: I1129 04:42:33.359656 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 29 04:42:33 crc kubenswrapper[4799]: I1129 04:42:33.410600 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 29 04:42:33 crc kubenswrapper[4799]: I1129 04:42:33.460369 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 29 04:42:33 crc kubenswrapper[4799]: I1129 04:42:33.558912 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 29 04:42:33 crc kubenswrapper[4799]: I1129 04:42:33.584925 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 29 04:42:33 crc kubenswrapper[4799]: I1129 04:42:33.601431 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 29 04:42:33 crc kubenswrapper[4799]: I1129 04:42:33.605502 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 29 04:42:33 crc kubenswrapper[4799]: I1129 04:42:33.635808 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 29 04:42:33 crc kubenswrapper[4799]: I1129 04:42:33.669040 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 29 04:42:33 crc kubenswrapper[4799]: I1129 04:42:33.676124 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 29 04:42:33 crc kubenswrapper[4799]: I1129 04:42:33.715589 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 29 04:42:33 crc kubenswrapper[4799]: I1129 04:42:33.769155 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 29 04:42:33 crc kubenswrapper[4799]: I1129 04:42:33.842605 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 29 04:42:33 crc kubenswrapper[4799]: I1129 04:42:33.845136 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 29 04:42:33 crc kubenswrapper[4799]: I1129 04:42:33.972377 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 29 04:42:33 crc kubenswrapper[4799]: I1129 04:42:33.999783 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 29 04:42:34 crc kubenswrapper[4799]: I1129 04:42:34.064997 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 29 04:42:34 crc kubenswrapper[4799]: I1129 04:42:34.078108 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 29 04:42:34 crc kubenswrapper[4799]: I1129 04:42:34.183852 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 29 04:42:34 crc kubenswrapper[4799]: I1129 04:42:34.221832 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 29 04:42:34 crc kubenswrapper[4799]: I1129 04:42:34.241580 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 29 04:42:34 crc kubenswrapper[4799]: I1129 04:42:34.290031 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 29 04:42:34 crc kubenswrapper[4799]: I1129 04:42:34.357530 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 29 04:42:34 crc kubenswrapper[4799]: I1129 04:42:34.393024 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 29 04:42:34 crc kubenswrapper[4799]: I1129 04:42:34.526420 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 29 04:42:34 crc kubenswrapper[4799]: I1129 04:42:34.543358 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 29 04:42:34 crc kubenswrapper[4799]: I1129 04:42:34.547164 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 29 04:42:34 crc kubenswrapper[4799]: I1129 04:42:34.557150 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 29 04:42:34 crc kubenswrapper[4799]: I1129 04:42:34.559761 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 29 04:42:34 crc kubenswrapper[4799]: I1129 04:42:34.577584 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 29 04:42:34 crc kubenswrapper[4799]: I1129 04:42:34.666260 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 29 04:42:34 crc kubenswrapper[4799]: I1129 04:42:34.712178 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 29 04:42:34 crc kubenswrapper[4799]: I1129 04:42:34.718695 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 29 04:42:34 crc kubenswrapper[4799]: I1129 04:42:34.743113 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 29 04:42:34 crc kubenswrapper[4799]: I1129 04:42:34.830257 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 29 04:42:34 crc kubenswrapper[4799]: I1129 04:42:34.848852 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 29 04:42:34 crc kubenswrapper[4799]: I1129 04:42:34.856946 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 29 04:42:34 crc kubenswrapper[4799]: I1129 04:42:34.888043 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 29 04:42:34 crc kubenswrapper[4799]: I1129 04:42:34.923450 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 29 04:42:34 crc kubenswrapper[4799]: I1129 04:42:34.969936 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 29 04:42:35 crc kubenswrapper[4799]: I1129 04:42:35.063216 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 29 04:42:35 crc kubenswrapper[4799]: I1129 04:42:35.072221 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 29 04:42:35 crc kubenswrapper[4799]: I1129 04:42:35.163119 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 29 04:42:35 crc kubenswrapper[4799]: I1129 04:42:35.196525 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 29 04:42:35 crc kubenswrapper[4799]: I1129 04:42:35.288122 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 29 04:42:35 crc kubenswrapper[4799]: I1129 04:42:35.356912 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 29 04:42:35 crc kubenswrapper[4799]: I1129 04:42:35.389596 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 29 04:42:35 crc kubenswrapper[4799]: I1129 04:42:35.442472 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 29 04:42:35 crc kubenswrapper[4799]: I1129 04:42:35.490473 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 29 04:42:35 crc kubenswrapper[4799]: I1129 04:42:35.514722 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 29 04:42:35 crc kubenswrapper[4799]: I1129 04:42:35.524513 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 29 04:42:35 crc kubenswrapper[4799]: I1129 04:42:35.537134 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 29 04:42:35 crc kubenswrapper[4799]: I1129 04:42:35.548662 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 29 04:42:35 crc kubenswrapper[4799]: I1129 04:42:35.578922 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 29 04:42:35 crc kubenswrapper[4799]: I1129 04:42:35.629908 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 29 04:42:35 crc kubenswrapper[4799]: I1129 04:42:35.674822 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 29 04:42:35 crc kubenswrapper[4799]: I1129 04:42:35.773089 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 29 04:42:35 crc kubenswrapper[4799]: I1129 04:42:35.816827 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 29 04:42:35 crc kubenswrapper[4799]: I1129 04:42:35.869718 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 29 04:42:35 crc kubenswrapper[4799]: I1129 04:42:35.882301 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 29 04:42:35 crc kubenswrapper[4799]: I1129 04:42:35.997684 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 29 04:42:36 crc kubenswrapper[4799]: I1129 04:42:36.010221 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 29 04:42:36 crc kubenswrapper[4799]: I1129 04:42:36.015476 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 29 04:42:36 crc kubenswrapper[4799]: I1129 04:42:36.029445 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 29 04:42:36 crc kubenswrapper[4799]: I1129 04:42:36.041366 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 29 04:42:36 crc kubenswrapper[4799]: I1129 04:42:36.113779 4799 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 29 04:42:36 crc kubenswrapper[4799]: I1129 04:42:36.115845 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=45.115820379 podStartE2EDuration="45.115820379s" podCreationTimestamp="2025-11-29 04:41:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:42:14.109063622 +0000 UTC m=+209.751994032" watchObservedRunningTime="2025-11-29 04:42:36.115820379 +0000 UTC m=+231.758750789" Nov 29 04:42:36 crc kubenswrapper[4799]: I1129 04:42:36.121019 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 29 04:42:36 crc kubenswrapper[4799]: I1129 04:42:36.121250 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 29 04:42:36 crc kubenswrapper[4799]: I1129 04:42:36.121784 4799 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4d40160e-075f-44b7-9d9d-3291e7774f27" Nov 29 04:42:36 crc kubenswrapper[4799]: I1129 04:42:36.121871 4799 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4d40160e-075f-44b7-9d9d-3291e7774f27" Nov 29 04:42:36 crc kubenswrapper[4799]: I1129 04:42:36.125738 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 04:42:36 crc kubenswrapper[4799]: I1129 04:42:36.145401 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=22.14538071 podStartE2EDuration="22.14538071s" podCreationTimestamp="2025-11-29 04:42:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:42:36.143471071 +0000 UTC m=+231.786401481" watchObservedRunningTime="2025-11-29 04:42:36.14538071 +0000 UTC m=+231.788311110" Nov 29 04:42:36 crc kubenswrapper[4799]: I1129 04:42:36.157298 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 29 04:42:36 crc kubenswrapper[4799]: I1129 04:42:36.178252 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 29 04:42:36 crc kubenswrapper[4799]: I1129 04:42:36.210331 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 29 04:42:36 crc kubenswrapper[4799]: I1129 04:42:36.229203 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 29 04:42:36 crc kubenswrapper[4799]: I1129 04:42:36.309821 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 29 04:42:36 crc kubenswrapper[4799]: I1129 04:42:36.368537 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 04:42:36 crc kubenswrapper[4799]: I1129 04:42:36.390708 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 29 04:42:36 crc kubenswrapper[4799]: I1129 04:42:36.430095 4799 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 29 04:42:36 crc kubenswrapper[4799]: I1129 04:42:36.465367 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 29 04:42:36 crc kubenswrapper[4799]: I1129 04:42:36.542149 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 29 04:42:36 crc kubenswrapper[4799]: I1129 04:42:36.580143 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 29 04:42:36 crc kubenswrapper[4799]: I1129 04:42:36.621306 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 29 04:42:36 crc kubenswrapper[4799]: I1129 04:42:36.670222 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 29 04:42:36 crc kubenswrapper[4799]: I1129 04:42:36.734843 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 29 04:42:36 crc kubenswrapper[4799]: I1129 04:42:36.739476 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 29 04:42:36 crc kubenswrapper[4799]: I1129 04:42:36.825150 4799 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 29 04:42:36 crc kubenswrapper[4799]: I1129 04:42:36.825492 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://71b4c7ac05233cd3efa579f07ad67706ced0921ba20e8ec60cb539dcde2a0ec7" gracePeriod=5 Nov 29 04:42:36 crc kubenswrapper[4799]: I1129 04:42:36.878589 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.214529 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.270828 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dnz56"] Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.271121 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dnz56" podUID="8734a259-eaa8-459e-b38d-e821e336a863" containerName="registry-server" containerID="cri-o://583b97c0ef0a8b6a73ba2cb895701337abd14f0f1c1188cb916a0b1ad750e5a6" gracePeriod=30 Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.278925 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jlvpn"] Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.279302 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jlvpn" podUID="2544bf9f-2bb8-4661-997c-9b25a96103d4" containerName="registry-server" containerID="cri-o://65ddff9b858f131e61390635199130e34b293bad0972d10e2d625b6a83beac00" gracePeriod=30 Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.293288 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rhzbg"] Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.293598 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-rhzbg" podUID="fecaea20-4e1a-4977-92e0-d0bd7c3037a2" containerName="marketplace-operator" containerID="cri-o://24926035b65c4efde5d76e8b70deb5e7d7dbd2e35d5c73226ad419f1280bdae9" gracePeriod=30 Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.307077 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q6m25"] Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.307739 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-q6m25" podUID="0a4df431-deee-45d6-9c07-7a1616a23d91" containerName="registry-server" containerID="cri-o://32979ecb7f9458352b758f3316566a687c8ca038c4f6cb66ccacf500f508bf82" gracePeriod=30 Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.319407 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-br2p6"] Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.319690 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-br2p6" podUID="390f5da3-eb7e-48af-bcc0-666c9172403e" containerName="registry-server" containerID="cri-o://d33a9c8f1300805eea0ebb98dbb69b5b8e6f188ed0ac37e0ce2c483a4c223675" gracePeriod=30 Nov 29 04:42:37 crc kubenswrapper[4799]: E1129 04:42:37.421175 4799 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfecaea20_4e1a_4977_92e0_d0bd7c3037a2.slice/crio-24926035b65c4efde5d76e8b70deb5e7d7dbd2e35d5c73226ad419f1280bdae9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8734a259_eaa8_459e_b38d_e821e336a863.slice/crio-conmon-583b97c0ef0a8b6a73ba2cb895701337abd14f0f1c1188cb916a0b1ad750e5a6.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8734a259_eaa8_459e_b38d_e821e336a863.slice/crio-583b97c0ef0a8b6a73ba2cb895701337abd14f0f1c1188cb916a0b1ad750e5a6.scope\": RecentStats: unable to find data in memory cache]" Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.429464 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-szfkm"] Nov 29 04:42:37 crc kubenswrapper[4799]: E1129 04:42:37.429779 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.429816 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 29 04:42:37 crc kubenswrapper[4799]: E1129 04:42:37.429841 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dae4078-40b8-46a7-a83c-d162a1295c27" containerName="installer" Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.429848 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dae4078-40b8-46a7-a83c-d162a1295c27" containerName="installer" Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.429974 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.429991 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="7dae4078-40b8-46a7-a83c-d162a1295c27" containerName="installer" Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.430499 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-szfkm" Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.444873 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-szfkm"] Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.479733 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.480053 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.499158 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.499230 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.499278 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-26t88" Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.500209 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"83b2ec8f0d384066ba90eb42c1a0a925ade3722b2515b0592d6f2a8f5e533a0f"} pod="openshift-machine-config-operator/machine-config-daemon-26t88" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.500278 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" containerID="cri-o://83b2ec8f0d384066ba90eb42c1a0a925ade3722b2515b0592d6f2a8f5e533a0f" gracePeriod=600 Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.519739 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.546347 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.576921 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6zrx\" (UniqueName: \"kubernetes.io/projected/72063360-3635-4677-8d2a-8d639af3ad6d-kube-api-access-h6zrx\") pod \"marketplace-operator-79b997595-szfkm\" (UID: \"72063360-3635-4677-8d2a-8d639af3ad6d\") " pod="openshift-marketplace/marketplace-operator-79b997595-szfkm" Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.577242 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/72063360-3635-4677-8d2a-8d639af3ad6d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-szfkm\" (UID: \"72063360-3635-4677-8d2a-8d639af3ad6d\") " pod="openshift-marketplace/marketplace-operator-79b997595-szfkm" Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.577308 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/72063360-3635-4677-8d2a-8d639af3ad6d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-szfkm\" (UID: \"72063360-3635-4677-8d2a-8d639af3ad6d\") " pod="openshift-marketplace/marketplace-operator-79b997595-szfkm" Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.679260 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/72063360-3635-4677-8d2a-8d639af3ad6d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-szfkm\" (UID: \"72063360-3635-4677-8d2a-8d639af3ad6d\") " pod="openshift-marketplace/marketplace-operator-79b997595-szfkm" Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.681987 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6zrx\" (UniqueName: \"kubernetes.io/projected/72063360-3635-4677-8d2a-8d639af3ad6d-kube-api-access-h6zrx\") pod \"marketplace-operator-79b997595-szfkm\" (UID: \"72063360-3635-4677-8d2a-8d639af3ad6d\") " pod="openshift-marketplace/marketplace-operator-79b997595-szfkm" Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.682044 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/72063360-3635-4677-8d2a-8d639af3ad6d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-szfkm\" (UID: \"72063360-3635-4677-8d2a-8d639af3ad6d\") " pod="openshift-marketplace/marketplace-operator-79b997595-szfkm" Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.682336 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/72063360-3635-4677-8d2a-8d639af3ad6d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-szfkm\" (UID: \"72063360-3635-4677-8d2a-8d639af3ad6d\") " pod="openshift-marketplace/marketplace-operator-79b997595-szfkm" Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.693450 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/72063360-3635-4677-8d2a-8d639af3ad6d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-szfkm\" (UID: \"72063360-3635-4677-8d2a-8d639af3ad6d\") " pod="openshift-marketplace/marketplace-operator-79b997595-szfkm" Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.703929 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6zrx\" (UniqueName: \"kubernetes.io/projected/72063360-3635-4677-8d2a-8d639af3ad6d-kube-api-access-h6zrx\") pod \"marketplace-operator-79b997595-szfkm\" (UID: \"72063360-3635-4677-8d2a-8d639af3ad6d\") " pod="openshift-marketplace/marketplace-operator-79b997595-szfkm" Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.774758 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-szfkm" Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.775462 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jlvpn" Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.776499 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.792312 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dnz56" Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.818405 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.870043 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.886692 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-br2p6" Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.888724 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2544bf9f-2bb8-4661-997c-9b25a96103d4-catalog-content\") pod \"2544bf9f-2bb8-4661-997c-9b25a96103d4\" (UID: \"2544bf9f-2bb8-4661-997c-9b25a96103d4\") " Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.888851 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6ssg\" (UniqueName: \"kubernetes.io/projected/2544bf9f-2bb8-4661-997c-9b25a96103d4-kube-api-access-n6ssg\") pod \"2544bf9f-2bb8-4661-997c-9b25a96103d4\" (UID: \"2544bf9f-2bb8-4661-997c-9b25a96103d4\") " Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.888911 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8734a259-eaa8-459e-b38d-e821e336a863-catalog-content\") pod \"8734a259-eaa8-459e-b38d-e821e336a863\" (UID: \"8734a259-eaa8-459e-b38d-e821e336a863\") " Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.888942 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b78n8\" (UniqueName: \"kubernetes.io/projected/8734a259-eaa8-459e-b38d-e821e336a863-kube-api-access-b78n8\") pod \"8734a259-eaa8-459e-b38d-e821e336a863\" (UID: \"8734a259-eaa8-459e-b38d-e821e336a863\") " Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.889041 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8734a259-eaa8-459e-b38d-e821e336a863-utilities\") pod \"8734a259-eaa8-459e-b38d-e821e336a863\" (UID: \"8734a259-eaa8-459e-b38d-e821e336a863\") " Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.889069 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2544bf9f-2bb8-4661-997c-9b25a96103d4-utilities\") pod \"2544bf9f-2bb8-4661-997c-9b25a96103d4\" (UID: \"2544bf9f-2bb8-4661-997c-9b25a96103d4\") " Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.892169 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8734a259-eaa8-459e-b38d-e821e336a863-utilities" (OuterVolumeSpecName: "utilities") pod "8734a259-eaa8-459e-b38d-e821e336a863" (UID: "8734a259-eaa8-459e-b38d-e821e336a863"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.892820 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2544bf9f-2bb8-4661-997c-9b25a96103d4-utilities" (OuterVolumeSpecName: "utilities") pod "2544bf9f-2bb8-4661-997c-9b25a96103d4" (UID: "2544bf9f-2bb8-4661-997c-9b25a96103d4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.897494 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2544bf9f-2bb8-4661-997c-9b25a96103d4-kube-api-access-n6ssg" (OuterVolumeSpecName: "kube-api-access-n6ssg") pod "2544bf9f-2bb8-4661-997c-9b25a96103d4" (UID: "2544bf9f-2bb8-4661-997c-9b25a96103d4"). InnerVolumeSpecName "kube-api-access-n6ssg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.897676 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8734a259-eaa8-459e-b38d-e821e336a863-kube-api-access-b78n8" (OuterVolumeSpecName: "kube-api-access-b78n8") pod "8734a259-eaa8-459e-b38d-e821e336a863" (UID: "8734a259-eaa8-459e-b38d-e821e336a863"). InnerVolumeSpecName "kube-api-access-b78n8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.899271 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-rhzbg" Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.907028 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q6m25" Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.937541 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.979956 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2544bf9f-2bb8-4661-997c-9b25a96103d4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2544bf9f-2bb8-4661-997c-9b25a96103d4" (UID: "2544bf9f-2bb8-4661-997c-9b25a96103d4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.986889 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8734a259-eaa8-459e-b38d-e821e336a863-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8734a259-eaa8-459e-b38d-e821e336a863" (UID: "8734a259-eaa8-459e-b38d-e821e336a863"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.989884 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/390f5da3-eb7e-48af-bcc0-666c9172403e-catalog-content\") pod \"390f5da3-eb7e-48af-bcc0-666c9172403e\" (UID: \"390f5da3-eb7e-48af-bcc0-666c9172403e\") " Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.989922 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzpbl\" (UniqueName: \"kubernetes.io/projected/fecaea20-4e1a-4977-92e0-d0bd7c3037a2-kube-api-access-fzpbl\") pod \"fecaea20-4e1a-4977-92e0-d0bd7c3037a2\" (UID: \"fecaea20-4e1a-4977-92e0-d0bd7c3037a2\") " Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.990108 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5b4sk\" (UniqueName: \"kubernetes.io/projected/390f5da3-eb7e-48af-bcc0-666c9172403e-kube-api-access-5b4sk\") pod \"390f5da3-eb7e-48af-bcc0-666c9172403e\" (UID: \"390f5da3-eb7e-48af-bcc0-666c9172403e\") " Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.990142 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/390f5da3-eb7e-48af-bcc0-666c9172403e-utilities\") pod \"390f5da3-eb7e-48af-bcc0-666c9172403e\" (UID: \"390f5da3-eb7e-48af-bcc0-666c9172403e\") " Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.990199 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fecaea20-4e1a-4977-92e0-d0bd7c3037a2-marketplace-trusted-ca\") pod \"fecaea20-4e1a-4977-92e0-d0bd7c3037a2\" (UID: \"fecaea20-4e1a-4977-92e0-d0bd7c3037a2\") " Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.990275 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/fecaea20-4e1a-4977-92e0-d0bd7c3037a2-marketplace-operator-metrics\") pod \"fecaea20-4e1a-4977-92e0-d0bd7c3037a2\" (UID: \"fecaea20-4e1a-4977-92e0-d0bd7c3037a2\") " Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.990525 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8734a259-eaa8-459e-b38d-e821e336a863-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.990541 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b78n8\" (UniqueName: \"kubernetes.io/projected/8734a259-eaa8-459e-b38d-e821e336a863-kube-api-access-b78n8\") on node \"crc\" DevicePath \"\"" Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.990557 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8734a259-eaa8-459e-b38d-e821e336a863-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.990568 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2544bf9f-2bb8-4661-997c-9b25a96103d4-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.990578 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2544bf9f-2bb8-4661-997c-9b25a96103d4-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.990587 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6ssg\" (UniqueName: \"kubernetes.io/projected/2544bf9f-2bb8-4661-997c-9b25a96103d4-kube-api-access-n6ssg\") on node \"crc\" DevicePath \"\"" Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.993458 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/390f5da3-eb7e-48af-bcc0-666c9172403e-utilities" (OuterVolumeSpecName: "utilities") pod "390f5da3-eb7e-48af-bcc0-666c9172403e" (UID: "390f5da3-eb7e-48af-bcc0-666c9172403e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:42:37 crc kubenswrapper[4799]: I1129 04:42:37.995169 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fecaea20-4e1a-4977-92e0-d0bd7c3037a2-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "fecaea20-4e1a-4977-92e0-d0bd7c3037a2" (UID: "fecaea20-4e1a-4977-92e0-d0bd7c3037a2"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.003613 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.004848 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/390f5da3-eb7e-48af-bcc0-666c9172403e-kube-api-access-5b4sk" (OuterVolumeSpecName: "kube-api-access-5b4sk") pod "390f5da3-eb7e-48af-bcc0-666c9172403e" (UID: "390f5da3-eb7e-48af-bcc0-666c9172403e"). InnerVolumeSpecName "kube-api-access-5b4sk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.006109 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fecaea20-4e1a-4977-92e0-d0bd7c3037a2-kube-api-access-fzpbl" (OuterVolumeSpecName: "kube-api-access-fzpbl") pod "fecaea20-4e1a-4977-92e0-d0bd7c3037a2" (UID: "fecaea20-4e1a-4977-92e0-d0bd7c3037a2"). InnerVolumeSpecName "kube-api-access-fzpbl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.006301 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fecaea20-4e1a-4977-92e0-d0bd7c3037a2-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "fecaea20-4e1a-4977-92e0-d0bd7c3037a2" (UID: "fecaea20-4e1a-4977-92e0-d0bd7c3037a2"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.014721 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.068906 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-szfkm"] Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.091735 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8p5ww\" (UniqueName: \"kubernetes.io/projected/0a4df431-deee-45d6-9c07-7a1616a23d91-kube-api-access-8p5ww\") pod \"0a4df431-deee-45d6-9c07-7a1616a23d91\" (UID: \"0a4df431-deee-45d6-9c07-7a1616a23d91\") " Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.092256 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a4df431-deee-45d6-9c07-7a1616a23d91-catalog-content\") pod \"0a4df431-deee-45d6-9c07-7a1616a23d91\" (UID: \"0a4df431-deee-45d6-9c07-7a1616a23d91\") " Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.096608 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a4df431-deee-45d6-9c07-7a1616a23d91-utilities\") pod \"0a4df431-deee-45d6-9c07-7a1616a23d91\" (UID: \"0a4df431-deee-45d6-9c07-7a1616a23d91\") " Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.095397 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a4df431-deee-45d6-9c07-7a1616a23d91-kube-api-access-8p5ww" (OuterVolumeSpecName: "kube-api-access-8p5ww") pod "0a4df431-deee-45d6-9c07-7a1616a23d91" (UID: "0a4df431-deee-45d6-9c07-7a1616a23d91"). InnerVolumeSpecName "kube-api-access-8p5ww". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.098196 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8p5ww\" (UniqueName: \"kubernetes.io/projected/0a4df431-deee-45d6-9c07-7a1616a23d91-kube-api-access-8p5ww\") on node \"crc\" DevicePath \"\"" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.098343 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5b4sk\" (UniqueName: \"kubernetes.io/projected/390f5da3-eb7e-48af-bcc0-666c9172403e-kube-api-access-5b4sk\") on node \"crc\" DevicePath \"\"" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.098509 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/390f5da3-eb7e-48af-bcc0-666c9172403e-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.098697 4799 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fecaea20-4e1a-4977-92e0-d0bd7c3037a2-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.098858 4799 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/fecaea20-4e1a-4977-92e0-d0bd7c3037a2-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.099081 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzpbl\" (UniqueName: \"kubernetes.io/projected/fecaea20-4e1a-4977-92e0-d0bd7c3037a2-kube-api-access-fzpbl\") on node \"crc\" DevicePath \"\"" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.108120 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a4df431-deee-45d6-9c07-7a1616a23d91-utilities" (OuterVolumeSpecName: "utilities") pod "0a4df431-deee-45d6-9c07-7a1616a23d91" (UID: "0a4df431-deee-45d6-9c07-7a1616a23d91"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.118431 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a4df431-deee-45d6-9c07-7a1616a23d91-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0a4df431-deee-45d6-9c07-7a1616a23d91" (UID: "0a4df431-deee-45d6-9c07-7a1616a23d91"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.128969 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/390f5da3-eb7e-48af-bcc0-666c9172403e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "390f5da3-eb7e-48af-bcc0-666c9172403e" (UID: "390f5da3-eb7e-48af-bcc0-666c9172403e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.163162 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.200096 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/390f5da3-eb7e-48af-bcc0-666c9172403e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.200616 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a4df431-deee-45d6-9c07-7a1616a23d91-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.200630 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a4df431-deee-45d6-9c07-7a1616a23d91-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.222815 4799 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.329940 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.376155 4799 generic.go:334] "Generic (PLEG): container finished" podID="2544bf9f-2bb8-4661-997c-9b25a96103d4" containerID="65ddff9b858f131e61390635199130e34b293bad0972d10e2d625b6a83beac00" exitCode=0 Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.376241 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jlvpn" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.376287 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jlvpn" event={"ID":"2544bf9f-2bb8-4661-997c-9b25a96103d4","Type":"ContainerDied","Data":"65ddff9b858f131e61390635199130e34b293bad0972d10e2d625b6a83beac00"} Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.376327 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jlvpn" event={"ID":"2544bf9f-2bb8-4661-997c-9b25a96103d4","Type":"ContainerDied","Data":"e02b6a5cc9c9ab9231d928449e29cf566f7b55db8a447c83cf25edae5434977b"} Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.376353 4799 scope.go:117] "RemoveContainer" containerID="65ddff9b858f131e61390635199130e34b293bad0972d10e2d625b6a83beac00" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.378314 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-szfkm" event={"ID":"72063360-3635-4677-8d2a-8d639af3ad6d","Type":"ContainerStarted","Data":"b9da5973ab25c9325a075686127ce158c62e31d926c1fb54773841860c8e95f5"} Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.378380 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-szfkm" event={"ID":"72063360-3635-4677-8d2a-8d639af3ad6d","Type":"ContainerStarted","Data":"61000f3a5eaf774c0b100f3e40c0f429ed8b03eb500c2deb3e79d57a1e668937"} Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.378896 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-szfkm" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.380120 4799 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-szfkm container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.57:8080/healthz\": dial tcp 10.217.0.57:8080: connect: connection refused" start-of-body= Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.380234 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-szfkm" podUID="72063360-3635-4677-8d2a-8d639af3ad6d" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.57:8080/healthz\": dial tcp 10.217.0.57:8080: connect: connection refused" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.380914 4799 generic.go:334] "Generic (PLEG): container finished" podID="390f5da3-eb7e-48af-bcc0-666c9172403e" containerID="d33a9c8f1300805eea0ebb98dbb69b5b8e6f188ed0ac37e0ce2c483a4c223675" exitCode=0 Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.380964 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-br2p6" event={"ID":"390f5da3-eb7e-48af-bcc0-666c9172403e","Type":"ContainerDied","Data":"d33a9c8f1300805eea0ebb98dbb69b5b8e6f188ed0ac37e0ce2c483a4c223675"} Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.380985 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-br2p6" event={"ID":"390f5da3-eb7e-48af-bcc0-666c9172403e","Type":"ContainerDied","Data":"07a5bc7f0abf4a702dd4d34fd6090cc33b820ce328d7c0a2854b6146fe98a8ab"} Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.381053 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-br2p6" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.387685 4799 generic.go:334] "Generic (PLEG): container finished" podID="8734a259-eaa8-459e-b38d-e821e336a863" containerID="583b97c0ef0a8b6a73ba2cb895701337abd14f0f1c1188cb916a0b1ad750e5a6" exitCode=0 Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.387895 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dnz56" event={"ID":"8734a259-eaa8-459e-b38d-e821e336a863","Type":"ContainerDied","Data":"583b97c0ef0a8b6a73ba2cb895701337abd14f0f1c1188cb916a0b1ad750e5a6"} Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.387985 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dnz56" event={"ID":"8734a259-eaa8-459e-b38d-e821e336a863","Type":"ContainerDied","Data":"5d6d52963f632c487a32a238c4563ce502479a251bf1b92c8118cb8184aec24d"} Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.388147 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dnz56" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.391750 4799 generic.go:334] "Generic (PLEG): container finished" podID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerID="83b2ec8f0d384066ba90eb42c1a0a925ade3722b2515b0592d6f2a8f5e533a0f" exitCode=0 Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.391848 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" event={"ID":"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8","Type":"ContainerDied","Data":"83b2ec8f0d384066ba90eb42c1a0a925ade3722b2515b0592d6f2a8f5e533a0f"} Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.391880 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" event={"ID":"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8","Type":"ContainerStarted","Data":"0ae025ef68b62241dcb107a107b30c8479869f4f96294d65af8cc7c29fb998ff"} Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.394282 4799 scope.go:117] "RemoveContainer" containerID="9f6fff8b5ce57b269a0dd398202181c3a76dcc3aa7486e709184bf12d934965e" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.395544 4799 generic.go:334] "Generic (PLEG): container finished" podID="0a4df431-deee-45d6-9c07-7a1616a23d91" containerID="32979ecb7f9458352b758f3316566a687c8ca038c4f6cb66ccacf500f508bf82" exitCode=0 Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.395599 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q6m25" event={"ID":"0a4df431-deee-45d6-9c07-7a1616a23d91","Type":"ContainerDied","Data":"32979ecb7f9458352b758f3316566a687c8ca038c4f6cb66ccacf500f508bf82"} Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.395625 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q6m25" event={"ID":"0a4df431-deee-45d6-9c07-7a1616a23d91","Type":"ContainerDied","Data":"18dca1cad7a736e26a87a4a2d9e1e92da1cff0cc3e1e220c2b42e8e26b2058ce"} Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.395692 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q6m25" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.404607 4799 generic.go:334] "Generic (PLEG): container finished" podID="fecaea20-4e1a-4977-92e0-d0bd7c3037a2" containerID="24926035b65c4efde5d76e8b70deb5e7d7dbd2e35d5c73226ad419f1280bdae9" exitCode=0 Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.405541 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-rhzbg" event={"ID":"fecaea20-4e1a-4977-92e0-d0bd7c3037a2","Type":"ContainerDied","Data":"24926035b65c4efde5d76e8b70deb5e7d7dbd2e35d5c73226ad419f1280bdae9"} Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.405609 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-rhzbg" event={"ID":"fecaea20-4e1a-4977-92e0-d0bd7c3037a2","Type":"ContainerDied","Data":"ba7cb2066e7577d7203ab116da03c50c95efc045b17699d2cbf3194bf3fa70a8"} Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.405634 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-rhzbg" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.418250 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-szfkm" podStartSLOduration=1.418214752 podStartE2EDuration="1.418214752s" podCreationTimestamp="2025-11-29 04:42:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:42:38.410782457 +0000 UTC m=+234.053712877" watchObservedRunningTime="2025-11-29 04:42:38.418214752 +0000 UTC m=+234.061145152" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.448047 4799 scope.go:117] "RemoveContainer" containerID="0915c197c63cf6faf6c69d1fa7d87570a8cc2cf97b65a0b4a4e8b99d67ec95d3" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.463651 4799 scope.go:117] "RemoveContainer" containerID="65ddff9b858f131e61390635199130e34b293bad0972d10e2d625b6a83beac00" Nov 29 04:42:38 crc kubenswrapper[4799]: E1129 04:42:38.464289 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65ddff9b858f131e61390635199130e34b293bad0972d10e2d625b6a83beac00\": container with ID starting with 65ddff9b858f131e61390635199130e34b293bad0972d10e2d625b6a83beac00 not found: ID does not exist" containerID="65ddff9b858f131e61390635199130e34b293bad0972d10e2d625b6a83beac00" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.464325 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65ddff9b858f131e61390635199130e34b293bad0972d10e2d625b6a83beac00"} err="failed to get container status \"65ddff9b858f131e61390635199130e34b293bad0972d10e2d625b6a83beac00\": rpc error: code = NotFound desc = could not find container \"65ddff9b858f131e61390635199130e34b293bad0972d10e2d625b6a83beac00\": container with ID starting with 65ddff9b858f131e61390635199130e34b293bad0972d10e2d625b6a83beac00 not found: ID does not exist" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.464353 4799 scope.go:117] "RemoveContainer" containerID="9f6fff8b5ce57b269a0dd398202181c3a76dcc3aa7486e709184bf12d934965e" Nov 29 04:42:38 crc kubenswrapper[4799]: E1129 04:42:38.464870 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f6fff8b5ce57b269a0dd398202181c3a76dcc3aa7486e709184bf12d934965e\": container with ID starting with 9f6fff8b5ce57b269a0dd398202181c3a76dcc3aa7486e709184bf12d934965e not found: ID does not exist" containerID="9f6fff8b5ce57b269a0dd398202181c3a76dcc3aa7486e709184bf12d934965e" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.465004 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f6fff8b5ce57b269a0dd398202181c3a76dcc3aa7486e709184bf12d934965e"} err="failed to get container status \"9f6fff8b5ce57b269a0dd398202181c3a76dcc3aa7486e709184bf12d934965e\": rpc error: code = NotFound desc = could not find container \"9f6fff8b5ce57b269a0dd398202181c3a76dcc3aa7486e709184bf12d934965e\": container with ID starting with 9f6fff8b5ce57b269a0dd398202181c3a76dcc3aa7486e709184bf12d934965e not found: ID does not exist" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.465137 4799 scope.go:117] "RemoveContainer" containerID="0915c197c63cf6faf6c69d1fa7d87570a8cc2cf97b65a0b4a4e8b99d67ec95d3" Nov 29 04:42:38 crc kubenswrapper[4799]: E1129 04:42:38.465928 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0915c197c63cf6faf6c69d1fa7d87570a8cc2cf97b65a0b4a4e8b99d67ec95d3\": container with ID starting with 0915c197c63cf6faf6c69d1fa7d87570a8cc2cf97b65a0b4a4e8b99d67ec95d3 not found: ID does not exist" containerID="0915c197c63cf6faf6c69d1fa7d87570a8cc2cf97b65a0b4a4e8b99d67ec95d3" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.465961 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0915c197c63cf6faf6c69d1fa7d87570a8cc2cf97b65a0b4a4e8b99d67ec95d3"} err="failed to get container status \"0915c197c63cf6faf6c69d1fa7d87570a8cc2cf97b65a0b4a4e8b99d67ec95d3\": rpc error: code = NotFound desc = could not find container \"0915c197c63cf6faf6c69d1fa7d87570a8cc2cf97b65a0b4a4e8b99d67ec95d3\": container with ID starting with 0915c197c63cf6faf6c69d1fa7d87570a8cc2cf97b65a0b4a4e8b99d67ec95d3 not found: ID does not exist" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.465976 4799 scope.go:117] "RemoveContainer" containerID="d33a9c8f1300805eea0ebb98dbb69b5b8e6f188ed0ac37e0ce2c483a4c223675" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.483625 4799 scope.go:117] "RemoveContainer" containerID="ddde27ed7104ddebb76cfaca165251f8783fd7de7a5a7b91cd177cdc191cca55" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.498634 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dnz56"] Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.499219 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.504443 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.508450 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dnz56"] Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.514368 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-br2p6"] Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.517458 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-br2p6"] Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.517610 4799 scope.go:117] "RemoveContainer" containerID="dcf896bbfd487aee632ced936832d741251434c5f231ea563fbf32d3084cad97" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.519525 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.522055 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rhzbg"] Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.524952 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rhzbg"] Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.534227 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jlvpn"] Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.536087 4799 scope.go:117] "RemoveContainer" containerID="d33a9c8f1300805eea0ebb98dbb69b5b8e6f188ed0ac37e0ce2c483a4c223675" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.536148 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jlvpn"] Nov 29 04:42:38 crc kubenswrapper[4799]: E1129 04:42:38.536904 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d33a9c8f1300805eea0ebb98dbb69b5b8e6f188ed0ac37e0ce2c483a4c223675\": container with ID starting with d33a9c8f1300805eea0ebb98dbb69b5b8e6f188ed0ac37e0ce2c483a4c223675 not found: ID does not exist" containerID="d33a9c8f1300805eea0ebb98dbb69b5b8e6f188ed0ac37e0ce2c483a4c223675" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.536953 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d33a9c8f1300805eea0ebb98dbb69b5b8e6f188ed0ac37e0ce2c483a4c223675"} err="failed to get container status \"d33a9c8f1300805eea0ebb98dbb69b5b8e6f188ed0ac37e0ce2c483a4c223675\": rpc error: code = NotFound desc = could not find container \"d33a9c8f1300805eea0ebb98dbb69b5b8e6f188ed0ac37e0ce2c483a4c223675\": container with ID starting with d33a9c8f1300805eea0ebb98dbb69b5b8e6f188ed0ac37e0ce2c483a4c223675 not found: ID does not exist" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.536987 4799 scope.go:117] "RemoveContainer" containerID="ddde27ed7104ddebb76cfaca165251f8783fd7de7a5a7b91cd177cdc191cca55" Nov 29 04:42:38 crc kubenswrapper[4799]: E1129 04:42:38.537320 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ddde27ed7104ddebb76cfaca165251f8783fd7de7a5a7b91cd177cdc191cca55\": container with ID starting with ddde27ed7104ddebb76cfaca165251f8783fd7de7a5a7b91cd177cdc191cca55 not found: ID does not exist" containerID="ddde27ed7104ddebb76cfaca165251f8783fd7de7a5a7b91cd177cdc191cca55" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.537351 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddde27ed7104ddebb76cfaca165251f8783fd7de7a5a7b91cd177cdc191cca55"} err="failed to get container status \"ddde27ed7104ddebb76cfaca165251f8783fd7de7a5a7b91cd177cdc191cca55\": rpc error: code = NotFound desc = could not find container \"ddde27ed7104ddebb76cfaca165251f8783fd7de7a5a7b91cd177cdc191cca55\": container with ID starting with ddde27ed7104ddebb76cfaca165251f8783fd7de7a5a7b91cd177cdc191cca55 not found: ID does not exist" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.537367 4799 scope.go:117] "RemoveContainer" containerID="dcf896bbfd487aee632ced936832d741251434c5f231ea563fbf32d3084cad97" Nov 29 04:42:38 crc kubenswrapper[4799]: E1129 04:42:38.537697 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dcf896bbfd487aee632ced936832d741251434c5f231ea563fbf32d3084cad97\": container with ID starting with dcf896bbfd487aee632ced936832d741251434c5f231ea563fbf32d3084cad97 not found: ID does not exist" containerID="dcf896bbfd487aee632ced936832d741251434c5f231ea563fbf32d3084cad97" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.537774 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dcf896bbfd487aee632ced936832d741251434c5f231ea563fbf32d3084cad97"} err="failed to get container status \"dcf896bbfd487aee632ced936832d741251434c5f231ea563fbf32d3084cad97\": rpc error: code = NotFound desc = could not find container \"dcf896bbfd487aee632ced936832d741251434c5f231ea563fbf32d3084cad97\": container with ID starting with dcf896bbfd487aee632ced936832d741251434c5f231ea563fbf32d3084cad97 not found: ID does not exist" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.537808 4799 scope.go:117] "RemoveContainer" containerID="583b97c0ef0a8b6a73ba2cb895701337abd14f0f1c1188cb916a0b1ad750e5a6" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.549449 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q6m25"] Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.554499 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-q6m25"] Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.557134 4799 scope.go:117] "RemoveContainer" containerID="728010b2c0c528c35fcfca45db3587f5630bf8ba9e9cad64ab24ca88cb02600e" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.577253 4799 scope.go:117] "RemoveContainer" containerID="8eeb2ca354b67247aa5a76641a69ba54701ee09f72d0a4f3ceaac0ac1e7903d1" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.594358 4799 scope.go:117] "RemoveContainer" containerID="583b97c0ef0a8b6a73ba2cb895701337abd14f0f1c1188cb916a0b1ad750e5a6" Nov 29 04:42:38 crc kubenswrapper[4799]: E1129 04:42:38.595049 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"583b97c0ef0a8b6a73ba2cb895701337abd14f0f1c1188cb916a0b1ad750e5a6\": container with ID starting with 583b97c0ef0a8b6a73ba2cb895701337abd14f0f1c1188cb916a0b1ad750e5a6 not found: ID does not exist" containerID="583b97c0ef0a8b6a73ba2cb895701337abd14f0f1c1188cb916a0b1ad750e5a6" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.595100 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"583b97c0ef0a8b6a73ba2cb895701337abd14f0f1c1188cb916a0b1ad750e5a6"} err="failed to get container status \"583b97c0ef0a8b6a73ba2cb895701337abd14f0f1c1188cb916a0b1ad750e5a6\": rpc error: code = NotFound desc = could not find container \"583b97c0ef0a8b6a73ba2cb895701337abd14f0f1c1188cb916a0b1ad750e5a6\": container with ID starting with 583b97c0ef0a8b6a73ba2cb895701337abd14f0f1c1188cb916a0b1ad750e5a6 not found: ID does not exist" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.595146 4799 scope.go:117] "RemoveContainer" containerID="728010b2c0c528c35fcfca45db3587f5630bf8ba9e9cad64ab24ca88cb02600e" Nov 29 04:42:38 crc kubenswrapper[4799]: E1129 04:42:38.595601 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"728010b2c0c528c35fcfca45db3587f5630bf8ba9e9cad64ab24ca88cb02600e\": container with ID starting with 728010b2c0c528c35fcfca45db3587f5630bf8ba9e9cad64ab24ca88cb02600e not found: ID does not exist" containerID="728010b2c0c528c35fcfca45db3587f5630bf8ba9e9cad64ab24ca88cb02600e" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.595659 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"728010b2c0c528c35fcfca45db3587f5630bf8ba9e9cad64ab24ca88cb02600e"} err="failed to get container status \"728010b2c0c528c35fcfca45db3587f5630bf8ba9e9cad64ab24ca88cb02600e\": rpc error: code = NotFound desc = could not find container \"728010b2c0c528c35fcfca45db3587f5630bf8ba9e9cad64ab24ca88cb02600e\": container with ID starting with 728010b2c0c528c35fcfca45db3587f5630bf8ba9e9cad64ab24ca88cb02600e not found: ID does not exist" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.595688 4799 scope.go:117] "RemoveContainer" containerID="8eeb2ca354b67247aa5a76641a69ba54701ee09f72d0a4f3ceaac0ac1e7903d1" Nov 29 04:42:38 crc kubenswrapper[4799]: E1129 04:42:38.596258 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8eeb2ca354b67247aa5a76641a69ba54701ee09f72d0a4f3ceaac0ac1e7903d1\": container with ID starting with 8eeb2ca354b67247aa5a76641a69ba54701ee09f72d0a4f3ceaac0ac1e7903d1 not found: ID does not exist" containerID="8eeb2ca354b67247aa5a76641a69ba54701ee09f72d0a4f3ceaac0ac1e7903d1" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.596290 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8eeb2ca354b67247aa5a76641a69ba54701ee09f72d0a4f3ceaac0ac1e7903d1"} err="failed to get container status \"8eeb2ca354b67247aa5a76641a69ba54701ee09f72d0a4f3ceaac0ac1e7903d1\": rpc error: code = NotFound desc = could not find container \"8eeb2ca354b67247aa5a76641a69ba54701ee09f72d0a4f3ceaac0ac1e7903d1\": container with ID starting with 8eeb2ca354b67247aa5a76641a69ba54701ee09f72d0a4f3ceaac0ac1e7903d1 not found: ID does not exist" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.596314 4799 scope.go:117] "RemoveContainer" containerID="32979ecb7f9458352b758f3316566a687c8ca038c4f6cb66ccacf500f508bf82" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.613351 4799 scope.go:117] "RemoveContainer" containerID="49a1c143a43a03d721de97043194fb50472a26cb8e36f0f35330fe7a8895f5a9" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.630128 4799 scope.go:117] "RemoveContainer" containerID="1ad6d883f0343277fb0d10110b42ad0e79a341d9e6e433adf083700d7d30b40a" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.645819 4799 scope.go:117] "RemoveContainer" containerID="32979ecb7f9458352b758f3316566a687c8ca038c4f6cb66ccacf500f508bf82" Nov 29 04:42:38 crc kubenswrapper[4799]: E1129 04:42:38.646354 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32979ecb7f9458352b758f3316566a687c8ca038c4f6cb66ccacf500f508bf82\": container with ID starting with 32979ecb7f9458352b758f3316566a687c8ca038c4f6cb66ccacf500f508bf82 not found: ID does not exist" containerID="32979ecb7f9458352b758f3316566a687c8ca038c4f6cb66ccacf500f508bf82" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.646389 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32979ecb7f9458352b758f3316566a687c8ca038c4f6cb66ccacf500f508bf82"} err="failed to get container status \"32979ecb7f9458352b758f3316566a687c8ca038c4f6cb66ccacf500f508bf82\": rpc error: code = NotFound desc = could not find container \"32979ecb7f9458352b758f3316566a687c8ca038c4f6cb66ccacf500f508bf82\": container with ID starting with 32979ecb7f9458352b758f3316566a687c8ca038c4f6cb66ccacf500f508bf82 not found: ID does not exist" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.646416 4799 scope.go:117] "RemoveContainer" containerID="49a1c143a43a03d721de97043194fb50472a26cb8e36f0f35330fe7a8895f5a9" Nov 29 04:42:38 crc kubenswrapper[4799]: E1129 04:42:38.646848 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49a1c143a43a03d721de97043194fb50472a26cb8e36f0f35330fe7a8895f5a9\": container with ID starting with 49a1c143a43a03d721de97043194fb50472a26cb8e36f0f35330fe7a8895f5a9 not found: ID does not exist" containerID="49a1c143a43a03d721de97043194fb50472a26cb8e36f0f35330fe7a8895f5a9" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.646874 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49a1c143a43a03d721de97043194fb50472a26cb8e36f0f35330fe7a8895f5a9"} err="failed to get container status \"49a1c143a43a03d721de97043194fb50472a26cb8e36f0f35330fe7a8895f5a9\": rpc error: code = NotFound desc = could not find container \"49a1c143a43a03d721de97043194fb50472a26cb8e36f0f35330fe7a8895f5a9\": container with ID starting with 49a1c143a43a03d721de97043194fb50472a26cb8e36f0f35330fe7a8895f5a9 not found: ID does not exist" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.646888 4799 scope.go:117] "RemoveContainer" containerID="1ad6d883f0343277fb0d10110b42ad0e79a341d9e6e433adf083700d7d30b40a" Nov 29 04:42:38 crc kubenswrapper[4799]: E1129 04:42:38.647217 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ad6d883f0343277fb0d10110b42ad0e79a341d9e6e433adf083700d7d30b40a\": container with ID starting with 1ad6d883f0343277fb0d10110b42ad0e79a341d9e6e433adf083700d7d30b40a not found: ID does not exist" containerID="1ad6d883f0343277fb0d10110b42ad0e79a341d9e6e433adf083700d7d30b40a" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.647239 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ad6d883f0343277fb0d10110b42ad0e79a341d9e6e433adf083700d7d30b40a"} err="failed to get container status \"1ad6d883f0343277fb0d10110b42ad0e79a341d9e6e433adf083700d7d30b40a\": rpc error: code = NotFound desc = could not find container \"1ad6d883f0343277fb0d10110b42ad0e79a341d9e6e433adf083700d7d30b40a\": container with ID starting with 1ad6d883f0343277fb0d10110b42ad0e79a341d9e6e433adf083700d7d30b40a not found: ID does not exist" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.647253 4799 scope.go:117] "RemoveContainer" containerID="24926035b65c4efde5d76e8b70deb5e7d7dbd2e35d5c73226ad419f1280bdae9" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.665107 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a4df431-deee-45d6-9c07-7a1616a23d91" path="/var/lib/kubelet/pods/0a4df431-deee-45d6-9c07-7a1616a23d91/volumes" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.665753 4799 scope.go:117] "RemoveContainer" containerID="24926035b65c4efde5d76e8b70deb5e7d7dbd2e35d5c73226ad419f1280bdae9" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.665850 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2544bf9f-2bb8-4661-997c-9b25a96103d4" path="/var/lib/kubelet/pods/2544bf9f-2bb8-4661-997c-9b25a96103d4/volumes" Nov 29 04:42:38 crc kubenswrapper[4799]: E1129 04:42:38.666114 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24926035b65c4efde5d76e8b70deb5e7d7dbd2e35d5c73226ad419f1280bdae9\": container with ID starting with 24926035b65c4efde5d76e8b70deb5e7d7dbd2e35d5c73226ad419f1280bdae9 not found: ID does not exist" containerID="24926035b65c4efde5d76e8b70deb5e7d7dbd2e35d5c73226ad419f1280bdae9" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.666175 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24926035b65c4efde5d76e8b70deb5e7d7dbd2e35d5c73226ad419f1280bdae9"} err="failed to get container status \"24926035b65c4efde5d76e8b70deb5e7d7dbd2e35d5c73226ad419f1280bdae9\": rpc error: code = NotFound desc = could not find container \"24926035b65c4efde5d76e8b70deb5e7d7dbd2e35d5c73226ad419f1280bdae9\": container with ID starting with 24926035b65c4efde5d76e8b70deb5e7d7dbd2e35d5c73226ad419f1280bdae9 not found: ID does not exist" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.666444 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="390f5da3-eb7e-48af-bcc0-666c9172403e" path="/var/lib/kubelet/pods/390f5da3-eb7e-48af-bcc0-666c9172403e/volumes" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.667572 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8734a259-eaa8-459e-b38d-e821e336a863" path="/var/lib/kubelet/pods/8734a259-eaa8-459e-b38d-e821e336a863/volumes" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.668251 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fecaea20-4e1a-4977-92e0-d0bd7c3037a2" path="/var/lib/kubelet/pods/fecaea20-4e1a-4977-92e0-d0bd7c3037a2/volumes" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.772145 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.883359 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.932573 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 29 04:42:38 crc kubenswrapper[4799]: I1129 04:42:38.936334 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 29 04:42:39 crc kubenswrapper[4799]: I1129 04:42:39.094432 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 29 04:42:39 crc kubenswrapper[4799]: I1129 04:42:39.096277 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 29 04:42:39 crc kubenswrapper[4799]: I1129 04:42:39.266834 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 29 04:42:39 crc kubenswrapper[4799]: I1129 04:42:39.325759 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 29 04:42:39 crc kubenswrapper[4799]: I1129 04:42:39.369130 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 29 04:42:39 crc kubenswrapper[4799]: I1129 04:42:39.386920 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 29 04:42:39 crc kubenswrapper[4799]: I1129 04:42:39.415641 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 29 04:42:39 crc kubenswrapper[4799]: I1129 04:42:39.423082 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-szfkm" Nov 29 04:42:39 crc kubenswrapper[4799]: I1129 04:42:39.448142 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 29 04:42:39 crc kubenswrapper[4799]: I1129 04:42:39.562538 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 29 04:42:39 crc kubenswrapper[4799]: I1129 04:42:39.604239 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 29 04:42:39 crc kubenswrapper[4799]: I1129 04:42:39.811970 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 29 04:42:40 crc kubenswrapper[4799]: I1129 04:42:40.045771 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 29 04:42:40 crc kubenswrapper[4799]: I1129 04:42:40.365496 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 29 04:42:42 crc kubenswrapper[4799]: I1129 04:42:42.404465 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 29 04:42:42 crc kubenswrapper[4799]: I1129 04:42:42.405016 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 29 04:42:42 crc kubenswrapper[4799]: I1129 04:42:42.435431 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 29 04:42:42 crc kubenswrapper[4799]: I1129 04:42:42.435513 4799 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="71b4c7ac05233cd3efa579f07ad67706ced0921ba20e8ec60cb539dcde2a0ec7" exitCode=137 Nov 29 04:42:42 crc kubenswrapper[4799]: I1129 04:42:42.435579 4799 scope.go:117] "RemoveContainer" containerID="71b4c7ac05233cd3efa579f07ad67706ced0921ba20e8ec60cb539dcde2a0ec7" Nov 29 04:42:42 crc kubenswrapper[4799]: I1129 04:42:42.435622 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 29 04:42:42 crc kubenswrapper[4799]: I1129 04:42:42.454851 4799 scope.go:117] "RemoveContainer" containerID="71b4c7ac05233cd3efa579f07ad67706ced0921ba20e8ec60cb539dcde2a0ec7" Nov 29 04:42:42 crc kubenswrapper[4799]: E1129 04:42:42.455239 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71b4c7ac05233cd3efa579f07ad67706ced0921ba20e8ec60cb539dcde2a0ec7\": container with ID starting with 71b4c7ac05233cd3efa579f07ad67706ced0921ba20e8ec60cb539dcde2a0ec7 not found: ID does not exist" containerID="71b4c7ac05233cd3efa579f07ad67706ced0921ba20e8ec60cb539dcde2a0ec7" Nov 29 04:42:42 crc kubenswrapper[4799]: I1129 04:42:42.455268 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71b4c7ac05233cd3efa579f07ad67706ced0921ba20e8ec60cb539dcde2a0ec7"} err="failed to get container status \"71b4c7ac05233cd3efa579f07ad67706ced0921ba20e8ec60cb539dcde2a0ec7\": rpc error: code = NotFound desc = could not find container \"71b4c7ac05233cd3efa579f07ad67706ced0921ba20e8ec60cb539dcde2a0ec7\": container with ID starting with 71b4c7ac05233cd3efa579f07ad67706ced0921ba20e8ec60cb539dcde2a0ec7 not found: ID does not exist" Nov 29 04:42:42 crc kubenswrapper[4799]: I1129 04:42:42.559594 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 29 04:42:42 crc kubenswrapper[4799]: I1129 04:42:42.559648 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 29 04:42:42 crc kubenswrapper[4799]: I1129 04:42:42.559678 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 29 04:42:42 crc kubenswrapper[4799]: I1129 04:42:42.559737 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 29 04:42:42 crc kubenswrapper[4799]: I1129 04:42:42.559860 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 04:42:42 crc kubenswrapper[4799]: I1129 04:42:42.559909 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 04:42:42 crc kubenswrapper[4799]: I1129 04:42:42.559928 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 04:42:42 crc kubenswrapper[4799]: I1129 04:42:42.559945 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 04:42:42 crc kubenswrapper[4799]: I1129 04:42:42.560866 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 29 04:42:42 crc kubenswrapper[4799]: I1129 04:42:42.561106 4799 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Nov 29 04:42:42 crc kubenswrapper[4799]: I1129 04:42:42.561144 4799 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Nov 29 04:42:42 crc kubenswrapper[4799]: I1129 04:42:42.561154 4799 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 29 04:42:42 crc kubenswrapper[4799]: I1129 04:42:42.561164 4799 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Nov 29 04:42:42 crc kubenswrapper[4799]: I1129 04:42:42.566558 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 04:42:42 crc kubenswrapper[4799]: I1129 04:42:42.661727 4799 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 29 04:42:42 crc kubenswrapper[4799]: I1129 04:42:42.665050 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Nov 29 04:42:42 crc kubenswrapper[4799]: I1129 04:42:42.665446 4799 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Nov 29 04:42:42 crc kubenswrapper[4799]: I1129 04:42:42.678917 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 29 04:42:42 crc kubenswrapper[4799]: I1129 04:42:42.678975 4799 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="b0ee9442-801d-4476-82fa-0dafcd1102f8" Nov 29 04:42:42 crc kubenswrapper[4799]: I1129 04:42:42.685739 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 29 04:42:42 crc kubenswrapper[4799]: I1129 04:42:42.685783 4799 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="b0ee9442-801d-4476-82fa-0dafcd1102f8" Nov 29 04:42:58 crc kubenswrapper[4799]: I1129 04:42:58.152346 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 29 04:42:59 crc kubenswrapper[4799]: I1129 04:42:59.786467 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-ddl6h"] Nov 29 04:42:59 crc kubenswrapper[4799]: I1129 04:42:59.786725 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-ddl6h" podUID="1c7126e3-85fe-4158-b92c-df2eb2717217" containerName="controller-manager" containerID="cri-o://a54d5d6a40e2a60ac476cb198cfdccb043ded7feb9925bd82143833a3786baed" gracePeriod=30 Nov 29 04:42:59 crc kubenswrapper[4799]: I1129 04:42:59.878053 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-z7df5"] Nov 29 04:42:59 crc kubenswrapper[4799]: I1129 04:42:59.878690 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-z7df5" podUID="a98eb357-7807-407c-a1e0-9d9b2b94e1fa" containerName="route-controller-manager" containerID="cri-o://eaaf73a2d39291bdd9d3c42cc35e90d2003d3cce6a94fc4dd29c313a9ae7b133" gracePeriod=30 Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.197485 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-ddl6h" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.282090 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-z7df5" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.330054 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c7126e3-85fe-4158-b92c-df2eb2717217-config\") pod \"1c7126e3-85fe-4158-b92c-df2eb2717217\" (UID: \"1c7126e3-85fe-4158-b92c-df2eb2717217\") " Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.330136 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8j7sv\" (UniqueName: \"kubernetes.io/projected/1c7126e3-85fe-4158-b92c-df2eb2717217-kube-api-access-8j7sv\") pod \"1c7126e3-85fe-4158-b92c-df2eb2717217\" (UID: \"1c7126e3-85fe-4158-b92c-df2eb2717217\") " Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.330264 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1c7126e3-85fe-4158-b92c-df2eb2717217-proxy-ca-bundles\") pod \"1c7126e3-85fe-4158-b92c-df2eb2717217\" (UID: \"1c7126e3-85fe-4158-b92c-df2eb2717217\") " Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.330660 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a98eb357-7807-407c-a1e0-9d9b2b94e1fa-client-ca\") pod \"a98eb357-7807-407c-a1e0-9d9b2b94e1fa\" (UID: \"a98eb357-7807-407c-a1e0-9d9b2b94e1fa\") " Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.330825 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvdj4\" (UniqueName: \"kubernetes.io/projected/a98eb357-7807-407c-a1e0-9d9b2b94e1fa-kube-api-access-kvdj4\") pod \"a98eb357-7807-407c-a1e0-9d9b2b94e1fa\" (UID: \"a98eb357-7807-407c-a1e0-9d9b2b94e1fa\") " Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.330919 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1c7126e3-85fe-4158-b92c-df2eb2717217-client-ca\") pod \"1c7126e3-85fe-4158-b92c-df2eb2717217\" (UID: \"1c7126e3-85fe-4158-b92c-df2eb2717217\") " Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.331060 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1c7126e3-85fe-4158-b92c-df2eb2717217-serving-cert\") pod \"1c7126e3-85fe-4158-b92c-df2eb2717217\" (UID: \"1c7126e3-85fe-4158-b92c-df2eb2717217\") " Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.331113 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a98eb357-7807-407c-a1e0-9d9b2b94e1fa-config\") pod \"a98eb357-7807-407c-a1e0-9d9b2b94e1fa\" (UID: \"a98eb357-7807-407c-a1e0-9d9b2b94e1fa\") " Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.331398 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c7126e3-85fe-4158-b92c-df2eb2717217-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "1c7126e3-85fe-4158-b92c-df2eb2717217" (UID: "1c7126e3-85fe-4158-b92c-df2eb2717217"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.331736 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c7126e3-85fe-4158-b92c-df2eb2717217-client-ca" (OuterVolumeSpecName: "client-ca") pod "1c7126e3-85fe-4158-b92c-df2eb2717217" (UID: "1c7126e3-85fe-4158-b92c-df2eb2717217"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.332205 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c7126e3-85fe-4158-b92c-df2eb2717217-config" (OuterVolumeSpecName: "config") pod "1c7126e3-85fe-4158-b92c-df2eb2717217" (UID: "1c7126e3-85fe-4158-b92c-df2eb2717217"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.332355 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a98eb357-7807-407c-a1e0-9d9b2b94e1fa-client-ca" (OuterVolumeSpecName: "client-ca") pod "a98eb357-7807-407c-a1e0-9d9b2b94e1fa" (UID: "a98eb357-7807-407c-a1e0-9d9b2b94e1fa"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.332534 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a98eb357-7807-407c-a1e0-9d9b2b94e1fa-config" (OuterVolumeSpecName: "config") pod "a98eb357-7807-407c-a1e0-9d9b2b94e1fa" (UID: "a98eb357-7807-407c-a1e0-9d9b2b94e1fa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.337390 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c7126e3-85fe-4158-b92c-df2eb2717217-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1c7126e3-85fe-4158-b92c-df2eb2717217" (UID: "1c7126e3-85fe-4158-b92c-df2eb2717217"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.339461 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a98eb357-7807-407c-a1e0-9d9b2b94e1fa-kube-api-access-kvdj4" (OuterVolumeSpecName: "kube-api-access-kvdj4") pod "a98eb357-7807-407c-a1e0-9d9b2b94e1fa" (UID: "a98eb357-7807-407c-a1e0-9d9b2b94e1fa"). InnerVolumeSpecName "kube-api-access-kvdj4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.340596 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c7126e3-85fe-4158-b92c-df2eb2717217-kube-api-access-8j7sv" (OuterVolumeSpecName: "kube-api-access-8j7sv") pod "1c7126e3-85fe-4158-b92c-df2eb2717217" (UID: "1c7126e3-85fe-4158-b92c-df2eb2717217"). InnerVolumeSpecName "kube-api-access-8j7sv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.388559 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7b6db79c5b-4bndf"] Nov 29 04:43:00 crc kubenswrapper[4799]: E1129 04:43:00.388918 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c7126e3-85fe-4158-b92c-df2eb2717217" containerName="controller-manager" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.388935 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c7126e3-85fe-4158-b92c-df2eb2717217" containerName="controller-manager" Nov 29 04:43:00 crc kubenswrapper[4799]: E1129 04:43:00.388947 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="390f5da3-eb7e-48af-bcc0-666c9172403e" containerName="registry-server" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.388975 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="390f5da3-eb7e-48af-bcc0-666c9172403e" containerName="registry-server" Nov 29 04:43:00 crc kubenswrapper[4799]: E1129 04:43:00.388985 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8734a259-eaa8-459e-b38d-e821e336a863" containerName="registry-server" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.388993 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="8734a259-eaa8-459e-b38d-e821e336a863" containerName="registry-server" Nov 29 04:43:00 crc kubenswrapper[4799]: E1129 04:43:00.389008 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2544bf9f-2bb8-4661-997c-9b25a96103d4" containerName="registry-server" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.389017 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="2544bf9f-2bb8-4661-997c-9b25a96103d4" containerName="registry-server" Nov 29 04:43:00 crc kubenswrapper[4799]: E1129 04:43:00.389029 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8734a259-eaa8-459e-b38d-e821e336a863" containerName="extract-utilities" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.389052 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="8734a259-eaa8-459e-b38d-e821e336a863" containerName="extract-utilities" Nov 29 04:43:00 crc kubenswrapper[4799]: E1129 04:43:00.389060 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a4df431-deee-45d6-9c07-7a1616a23d91" containerName="extract-content" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.389066 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a4df431-deee-45d6-9c07-7a1616a23d91" containerName="extract-content" Nov 29 04:43:00 crc kubenswrapper[4799]: E1129 04:43:00.389078 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2544bf9f-2bb8-4661-997c-9b25a96103d4" containerName="extract-utilities" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.389085 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="2544bf9f-2bb8-4661-997c-9b25a96103d4" containerName="extract-utilities" Nov 29 04:43:00 crc kubenswrapper[4799]: E1129 04:43:00.389095 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a4df431-deee-45d6-9c07-7a1616a23d91" containerName="registry-server" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.389103 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a4df431-deee-45d6-9c07-7a1616a23d91" containerName="registry-server" Nov 29 04:43:00 crc kubenswrapper[4799]: E1129 04:43:00.389132 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="390f5da3-eb7e-48af-bcc0-666c9172403e" containerName="extract-content" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.389141 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="390f5da3-eb7e-48af-bcc0-666c9172403e" containerName="extract-content" Nov 29 04:43:00 crc kubenswrapper[4799]: E1129 04:43:00.389149 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2544bf9f-2bb8-4661-997c-9b25a96103d4" containerName="extract-content" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.389157 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="2544bf9f-2bb8-4661-997c-9b25a96103d4" containerName="extract-content" Nov 29 04:43:00 crc kubenswrapper[4799]: E1129 04:43:00.389169 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a4df431-deee-45d6-9c07-7a1616a23d91" containerName="extract-utilities" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.389178 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a4df431-deee-45d6-9c07-7a1616a23d91" containerName="extract-utilities" Nov 29 04:43:00 crc kubenswrapper[4799]: E1129 04:43:00.389186 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="390f5da3-eb7e-48af-bcc0-666c9172403e" containerName="extract-utilities" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.389193 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="390f5da3-eb7e-48af-bcc0-666c9172403e" containerName="extract-utilities" Nov 29 04:43:00 crc kubenswrapper[4799]: E1129 04:43:00.389224 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a98eb357-7807-407c-a1e0-9d9b2b94e1fa" containerName="route-controller-manager" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.389233 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a98eb357-7807-407c-a1e0-9d9b2b94e1fa" containerName="route-controller-manager" Nov 29 04:43:00 crc kubenswrapper[4799]: E1129 04:43:00.389240 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8734a259-eaa8-459e-b38d-e821e336a863" containerName="extract-content" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.389247 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="8734a259-eaa8-459e-b38d-e821e336a863" containerName="extract-content" Nov 29 04:43:00 crc kubenswrapper[4799]: E1129 04:43:00.389261 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fecaea20-4e1a-4977-92e0-d0bd7c3037a2" containerName="marketplace-operator" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.389267 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="fecaea20-4e1a-4977-92e0-d0bd7c3037a2" containerName="marketplace-operator" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.389396 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="a98eb357-7807-407c-a1e0-9d9b2b94e1fa" containerName="route-controller-manager" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.389405 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="2544bf9f-2bb8-4661-997c-9b25a96103d4" containerName="registry-server" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.389413 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a4df431-deee-45d6-9c07-7a1616a23d91" containerName="registry-server" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.389425 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="390f5da3-eb7e-48af-bcc0-666c9172403e" containerName="registry-server" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.389448 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c7126e3-85fe-4158-b92c-df2eb2717217" containerName="controller-manager" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.389457 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="fecaea20-4e1a-4977-92e0-d0bd7c3037a2" containerName="marketplace-operator" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.389468 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="8734a259-eaa8-459e-b38d-e821e336a863" containerName="registry-server" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.389997 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7b6db79c5b-4bndf" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.399325 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7b6db79c5b-4bndf"] Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.415293 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-69cd86ff6b-chppz"] Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.416509 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-69cd86ff6b-chppz" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.431783 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a98eb357-7807-407c-a1e0-9d9b2b94e1fa-serving-cert\") pod \"a98eb357-7807-407c-a1e0-9d9b2b94e1fa\" (UID: \"a98eb357-7807-407c-a1e0-9d9b2b94e1fa\") " Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.432036 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5rqh\" (UniqueName: \"kubernetes.io/projected/55f389a3-e952-4cfb-8aec-4ae3aa563101-kube-api-access-v5rqh\") pod \"route-controller-manager-69cd86ff6b-chppz\" (UID: \"55f389a3-e952-4cfb-8aec-4ae3aa563101\") " pod="openshift-route-controller-manager/route-controller-manager-69cd86ff6b-chppz" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.432087 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55f389a3-e952-4cfb-8aec-4ae3aa563101-config\") pod \"route-controller-manager-69cd86ff6b-chppz\" (UID: \"55f389a3-e952-4cfb-8aec-4ae3aa563101\") " pod="openshift-route-controller-manager/route-controller-manager-69cd86ff6b-chppz" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.432116 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d10019fa-babc-4de4-be1b-df08d7247bc8-proxy-ca-bundles\") pod \"controller-manager-7b6db79c5b-4bndf\" (UID: \"d10019fa-babc-4de4-be1b-df08d7247bc8\") " pod="openshift-controller-manager/controller-manager-7b6db79c5b-4bndf" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.432142 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/55f389a3-e952-4cfb-8aec-4ae3aa563101-client-ca\") pod \"route-controller-manager-69cd86ff6b-chppz\" (UID: \"55f389a3-e952-4cfb-8aec-4ae3aa563101\") " pod="openshift-route-controller-manager/route-controller-manager-69cd86ff6b-chppz" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.432162 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d10019fa-babc-4de4-be1b-df08d7247bc8-client-ca\") pod \"controller-manager-7b6db79c5b-4bndf\" (UID: \"d10019fa-babc-4de4-be1b-df08d7247bc8\") " pod="openshift-controller-manager/controller-manager-7b6db79c5b-4bndf" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.432200 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d10019fa-babc-4de4-be1b-df08d7247bc8-config\") pod \"controller-manager-7b6db79c5b-4bndf\" (UID: \"d10019fa-babc-4de4-be1b-df08d7247bc8\") " pod="openshift-controller-manager/controller-manager-7b6db79c5b-4bndf" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.432247 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/55f389a3-e952-4cfb-8aec-4ae3aa563101-serving-cert\") pod \"route-controller-manager-69cd86ff6b-chppz\" (UID: \"55f389a3-e952-4cfb-8aec-4ae3aa563101\") " pod="openshift-route-controller-manager/route-controller-manager-69cd86ff6b-chppz" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.432290 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7f96h\" (UniqueName: \"kubernetes.io/projected/d10019fa-babc-4de4-be1b-df08d7247bc8-kube-api-access-7f96h\") pod \"controller-manager-7b6db79c5b-4bndf\" (UID: \"d10019fa-babc-4de4-be1b-df08d7247bc8\") " pod="openshift-controller-manager/controller-manager-7b6db79c5b-4bndf" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.432324 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d10019fa-babc-4de4-be1b-df08d7247bc8-serving-cert\") pod \"controller-manager-7b6db79c5b-4bndf\" (UID: \"d10019fa-babc-4de4-be1b-df08d7247bc8\") " pod="openshift-controller-manager/controller-manager-7b6db79c5b-4bndf" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.432377 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1c7126e3-85fe-4158-b92c-df2eb2717217-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.432394 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a98eb357-7807-407c-a1e0-9d9b2b94e1fa-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.432406 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8j7sv\" (UniqueName: \"kubernetes.io/projected/1c7126e3-85fe-4158-b92c-df2eb2717217-kube-api-access-8j7sv\") on node \"crc\" DevicePath \"\"" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.432419 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c7126e3-85fe-4158-b92c-df2eb2717217-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.432430 4799 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1c7126e3-85fe-4158-b92c-df2eb2717217-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.432444 4799 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a98eb357-7807-407c-a1e0-9d9b2b94e1fa-client-ca\") on node \"crc\" DevicePath \"\"" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.432457 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvdj4\" (UniqueName: \"kubernetes.io/projected/a98eb357-7807-407c-a1e0-9d9b2b94e1fa-kube-api-access-kvdj4\") on node \"crc\" DevicePath \"\"" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.432469 4799 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1c7126e3-85fe-4158-b92c-df2eb2717217-client-ca\") on node \"crc\" DevicePath \"\"" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.435958 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a98eb357-7807-407c-a1e0-9d9b2b94e1fa-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "a98eb357-7807-407c-a1e0-9d9b2b94e1fa" (UID: "a98eb357-7807-407c-a1e0-9d9b2b94e1fa"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.436303 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-69cd86ff6b-chppz"] Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.494844 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7b6db79c5b-4bndf"] Nov 29 04:43:00 crc kubenswrapper[4799]: E1129 04:43:00.495743 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[client-ca config kube-api-access-7f96h proxy-ca-bundles serving-cert], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openshift-controller-manager/controller-manager-7b6db79c5b-4bndf" podUID="d10019fa-babc-4de4-be1b-df08d7247bc8" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.533532 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7f96h\" (UniqueName: \"kubernetes.io/projected/d10019fa-babc-4de4-be1b-df08d7247bc8-kube-api-access-7f96h\") pod \"controller-manager-7b6db79c5b-4bndf\" (UID: \"d10019fa-babc-4de4-be1b-df08d7247bc8\") " pod="openshift-controller-manager/controller-manager-7b6db79c5b-4bndf" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.533602 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d10019fa-babc-4de4-be1b-df08d7247bc8-serving-cert\") pod \"controller-manager-7b6db79c5b-4bndf\" (UID: \"d10019fa-babc-4de4-be1b-df08d7247bc8\") " pod="openshift-controller-manager/controller-manager-7b6db79c5b-4bndf" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.533646 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5rqh\" (UniqueName: \"kubernetes.io/projected/55f389a3-e952-4cfb-8aec-4ae3aa563101-kube-api-access-v5rqh\") pod \"route-controller-manager-69cd86ff6b-chppz\" (UID: \"55f389a3-e952-4cfb-8aec-4ae3aa563101\") " pod="openshift-route-controller-manager/route-controller-manager-69cd86ff6b-chppz" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.533683 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55f389a3-e952-4cfb-8aec-4ae3aa563101-config\") pod \"route-controller-manager-69cd86ff6b-chppz\" (UID: \"55f389a3-e952-4cfb-8aec-4ae3aa563101\") " pod="openshift-route-controller-manager/route-controller-manager-69cd86ff6b-chppz" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.533708 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d10019fa-babc-4de4-be1b-df08d7247bc8-proxy-ca-bundles\") pod \"controller-manager-7b6db79c5b-4bndf\" (UID: \"d10019fa-babc-4de4-be1b-df08d7247bc8\") " pod="openshift-controller-manager/controller-manager-7b6db79c5b-4bndf" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.533734 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/55f389a3-e952-4cfb-8aec-4ae3aa563101-client-ca\") pod \"route-controller-manager-69cd86ff6b-chppz\" (UID: \"55f389a3-e952-4cfb-8aec-4ae3aa563101\") " pod="openshift-route-controller-manager/route-controller-manager-69cd86ff6b-chppz" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.533757 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d10019fa-babc-4de4-be1b-df08d7247bc8-client-ca\") pod \"controller-manager-7b6db79c5b-4bndf\" (UID: \"d10019fa-babc-4de4-be1b-df08d7247bc8\") " pod="openshift-controller-manager/controller-manager-7b6db79c5b-4bndf" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.533781 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d10019fa-babc-4de4-be1b-df08d7247bc8-config\") pod \"controller-manager-7b6db79c5b-4bndf\" (UID: \"d10019fa-babc-4de4-be1b-df08d7247bc8\") " pod="openshift-controller-manager/controller-manager-7b6db79c5b-4bndf" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.533845 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/55f389a3-e952-4cfb-8aec-4ae3aa563101-serving-cert\") pod \"route-controller-manager-69cd86ff6b-chppz\" (UID: \"55f389a3-e952-4cfb-8aec-4ae3aa563101\") " pod="openshift-route-controller-manager/route-controller-manager-69cd86ff6b-chppz" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.533909 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a98eb357-7807-407c-a1e0-9d9b2b94e1fa-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.535383 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d10019fa-babc-4de4-be1b-df08d7247bc8-client-ca\") pod \"controller-manager-7b6db79c5b-4bndf\" (UID: \"d10019fa-babc-4de4-be1b-df08d7247bc8\") " pod="openshift-controller-manager/controller-manager-7b6db79c5b-4bndf" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.535668 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/55f389a3-e952-4cfb-8aec-4ae3aa563101-client-ca\") pod \"route-controller-manager-69cd86ff6b-chppz\" (UID: \"55f389a3-e952-4cfb-8aec-4ae3aa563101\") " pod="openshift-route-controller-manager/route-controller-manager-69cd86ff6b-chppz" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.535692 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d10019fa-babc-4de4-be1b-df08d7247bc8-proxy-ca-bundles\") pod \"controller-manager-7b6db79c5b-4bndf\" (UID: \"d10019fa-babc-4de4-be1b-df08d7247bc8\") " pod="openshift-controller-manager/controller-manager-7b6db79c5b-4bndf" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.535778 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55f389a3-e952-4cfb-8aec-4ae3aa563101-config\") pod \"route-controller-manager-69cd86ff6b-chppz\" (UID: \"55f389a3-e952-4cfb-8aec-4ae3aa563101\") " pod="openshift-route-controller-manager/route-controller-manager-69cd86ff6b-chppz" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.536002 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d10019fa-babc-4de4-be1b-df08d7247bc8-config\") pod \"controller-manager-7b6db79c5b-4bndf\" (UID: \"d10019fa-babc-4de4-be1b-df08d7247bc8\") " pod="openshift-controller-manager/controller-manager-7b6db79c5b-4bndf" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.539267 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d10019fa-babc-4de4-be1b-df08d7247bc8-serving-cert\") pod \"controller-manager-7b6db79c5b-4bndf\" (UID: \"d10019fa-babc-4de4-be1b-df08d7247bc8\") " pod="openshift-controller-manager/controller-manager-7b6db79c5b-4bndf" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.539441 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/55f389a3-e952-4cfb-8aec-4ae3aa563101-serving-cert\") pod \"route-controller-manager-69cd86ff6b-chppz\" (UID: \"55f389a3-e952-4cfb-8aec-4ae3aa563101\") " pod="openshift-route-controller-manager/route-controller-manager-69cd86ff6b-chppz" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.551388 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7f96h\" (UniqueName: \"kubernetes.io/projected/d10019fa-babc-4de4-be1b-df08d7247bc8-kube-api-access-7f96h\") pod \"controller-manager-7b6db79c5b-4bndf\" (UID: \"d10019fa-babc-4de4-be1b-df08d7247bc8\") " pod="openshift-controller-manager/controller-manager-7b6db79c5b-4bndf" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.551659 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5rqh\" (UniqueName: \"kubernetes.io/projected/55f389a3-e952-4cfb-8aec-4ae3aa563101-kube-api-access-v5rqh\") pod \"route-controller-manager-69cd86ff6b-chppz\" (UID: \"55f389a3-e952-4cfb-8aec-4ae3aa563101\") " pod="openshift-route-controller-manager/route-controller-manager-69cd86ff6b-chppz" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.551705 4799 generic.go:334] "Generic (PLEG): container finished" podID="1c7126e3-85fe-4158-b92c-df2eb2717217" containerID="a54d5d6a40e2a60ac476cb198cfdccb043ded7feb9925bd82143833a3786baed" exitCode=0 Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.551777 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-ddl6h" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.551850 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-ddl6h" event={"ID":"1c7126e3-85fe-4158-b92c-df2eb2717217","Type":"ContainerDied","Data":"a54d5d6a40e2a60ac476cb198cfdccb043ded7feb9925bd82143833a3786baed"} Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.551888 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-ddl6h" event={"ID":"1c7126e3-85fe-4158-b92c-df2eb2717217","Type":"ContainerDied","Data":"d81239a6dc2d3950eeaf981a7e488978b8cc60f1af5358190a59f24c87edcd98"} Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.551909 4799 scope.go:117] "RemoveContainer" containerID="a54d5d6a40e2a60ac476cb198cfdccb043ded7feb9925bd82143833a3786baed" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.553180 4799 generic.go:334] "Generic (PLEG): container finished" podID="a98eb357-7807-407c-a1e0-9d9b2b94e1fa" containerID="eaaf73a2d39291bdd9d3c42cc35e90d2003d3cce6a94fc4dd29c313a9ae7b133" exitCode=0 Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.553229 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7b6db79c5b-4bndf" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.553612 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-z7df5" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.556336 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-z7df5" event={"ID":"a98eb357-7807-407c-a1e0-9d9b2b94e1fa","Type":"ContainerDied","Data":"eaaf73a2d39291bdd9d3c42cc35e90d2003d3cce6a94fc4dd29c313a9ae7b133"} Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.556400 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-z7df5" event={"ID":"a98eb357-7807-407c-a1e0-9d9b2b94e1fa","Type":"ContainerDied","Data":"80ed6fbe5b745575492889847c82f43bf37fd9315ccbb3b69bcb86b8e12d542a"} Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.563964 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7b6db79c5b-4bndf" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.576675 4799 scope.go:117] "RemoveContainer" containerID="a54d5d6a40e2a60ac476cb198cfdccb043ded7feb9925bd82143833a3786baed" Nov 29 04:43:00 crc kubenswrapper[4799]: E1129 04:43:00.577459 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a54d5d6a40e2a60ac476cb198cfdccb043ded7feb9925bd82143833a3786baed\": container with ID starting with a54d5d6a40e2a60ac476cb198cfdccb043ded7feb9925bd82143833a3786baed not found: ID does not exist" containerID="a54d5d6a40e2a60ac476cb198cfdccb043ded7feb9925bd82143833a3786baed" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.577541 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a54d5d6a40e2a60ac476cb198cfdccb043ded7feb9925bd82143833a3786baed"} err="failed to get container status \"a54d5d6a40e2a60ac476cb198cfdccb043ded7feb9925bd82143833a3786baed\": rpc error: code = NotFound desc = could not find container \"a54d5d6a40e2a60ac476cb198cfdccb043ded7feb9925bd82143833a3786baed\": container with ID starting with a54d5d6a40e2a60ac476cb198cfdccb043ded7feb9925bd82143833a3786baed not found: ID does not exist" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.577606 4799 scope.go:117] "RemoveContainer" containerID="eaaf73a2d39291bdd9d3c42cc35e90d2003d3cce6a94fc4dd29c313a9ae7b133" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.587768 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-ddl6h"] Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.591101 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-ddl6h"] Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.607214 4799 scope.go:117] "RemoveContainer" containerID="eaaf73a2d39291bdd9d3c42cc35e90d2003d3cce6a94fc4dd29c313a9ae7b133" Nov 29 04:43:00 crc kubenswrapper[4799]: E1129 04:43:00.610432 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eaaf73a2d39291bdd9d3c42cc35e90d2003d3cce6a94fc4dd29c313a9ae7b133\": container with ID starting with eaaf73a2d39291bdd9d3c42cc35e90d2003d3cce6a94fc4dd29c313a9ae7b133 not found: ID does not exist" containerID="eaaf73a2d39291bdd9d3c42cc35e90d2003d3cce6a94fc4dd29c313a9ae7b133" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.610488 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eaaf73a2d39291bdd9d3c42cc35e90d2003d3cce6a94fc4dd29c313a9ae7b133"} err="failed to get container status \"eaaf73a2d39291bdd9d3c42cc35e90d2003d3cce6a94fc4dd29c313a9ae7b133\": rpc error: code = NotFound desc = could not find container \"eaaf73a2d39291bdd9d3c42cc35e90d2003d3cce6a94fc4dd29c313a9ae7b133\": container with ID starting with eaaf73a2d39291bdd9d3c42cc35e90d2003d3cce6a94fc4dd29c313a9ae7b133 not found: ID does not exist" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.616967 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-z7df5"] Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.621144 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-z7df5"] Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.634733 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d10019fa-babc-4de4-be1b-df08d7247bc8-config\") pod \"d10019fa-babc-4de4-be1b-df08d7247bc8\" (UID: \"d10019fa-babc-4de4-be1b-df08d7247bc8\") " Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.634822 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d10019fa-babc-4de4-be1b-df08d7247bc8-client-ca\") pod \"d10019fa-babc-4de4-be1b-df08d7247bc8\" (UID: \"d10019fa-babc-4de4-be1b-df08d7247bc8\") " Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.634847 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d10019fa-babc-4de4-be1b-df08d7247bc8-proxy-ca-bundles\") pod \"d10019fa-babc-4de4-be1b-df08d7247bc8\" (UID: \"d10019fa-babc-4de4-be1b-df08d7247bc8\") " Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.634945 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7f96h\" (UniqueName: \"kubernetes.io/projected/d10019fa-babc-4de4-be1b-df08d7247bc8-kube-api-access-7f96h\") pod \"d10019fa-babc-4de4-be1b-df08d7247bc8\" (UID: \"d10019fa-babc-4de4-be1b-df08d7247bc8\") " Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.634968 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d10019fa-babc-4de4-be1b-df08d7247bc8-serving-cert\") pod \"d10019fa-babc-4de4-be1b-df08d7247bc8\" (UID: \"d10019fa-babc-4de4-be1b-df08d7247bc8\") " Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.635598 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d10019fa-babc-4de4-be1b-df08d7247bc8-client-ca" (OuterVolumeSpecName: "client-ca") pod "d10019fa-babc-4de4-be1b-df08d7247bc8" (UID: "d10019fa-babc-4de4-be1b-df08d7247bc8"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.635637 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d10019fa-babc-4de4-be1b-df08d7247bc8-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "d10019fa-babc-4de4-be1b-df08d7247bc8" (UID: "d10019fa-babc-4de4-be1b-df08d7247bc8"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.635713 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d10019fa-babc-4de4-be1b-df08d7247bc8-config" (OuterVolumeSpecName: "config") pod "d10019fa-babc-4de4-be1b-df08d7247bc8" (UID: "d10019fa-babc-4de4-be1b-df08d7247bc8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.639935 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d10019fa-babc-4de4-be1b-df08d7247bc8-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d10019fa-babc-4de4-be1b-df08d7247bc8" (UID: "d10019fa-babc-4de4-be1b-df08d7247bc8"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.640352 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d10019fa-babc-4de4-be1b-df08d7247bc8-kube-api-access-7f96h" (OuterVolumeSpecName: "kube-api-access-7f96h") pod "d10019fa-babc-4de4-be1b-df08d7247bc8" (UID: "d10019fa-babc-4de4-be1b-df08d7247bc8"). InnerVolumeSpecName "kube-api-access-7f96h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.666687 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c7126e3-85fe-4158-b92c-df2eb2717217" path="/var/lib/kubelet/pods/1c7126e3-85fe-4158-b92c-df2eb2717217/volumes" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.667471 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a98eb357-7807-407c-a1e0-9d9b2b94e1fa" path="/var/lib/kubelet/pods/a98eb357-7807-407c-a1e0-9d9b2b94e1fa/volumes" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.736156 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d10019fa-babc-4de4-be1b-df08d7247bc8-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.736207 4799 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d10019fa-babc-4de4-be1b-df08d7247bc8-client-ca\") on node \"crc\" DevicePath \"\"" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.736222 4799 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d10019fa-babc-4de4-be1b-df08d7247bc8-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.736238 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7f96h\" (UniqueName: \"kubernetes.io/projected/d10019fa-babc-4de4-be1b-df08d7247bc8-kube-api-access-7f96h\") on node \"crc\" DevicePath \"\"" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.736251 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d10019fa-babc-4de4-be1b-df08d7247bc8-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 04:43:00 crc kubenswrapper[4799]: I1129 04:43:00.736925 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-69cd86ff6b-chppz" Nov 29 04:43:01 crc kubenswrapper[4799]: I1129 04:43:01.026764 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-69cd86ff6b-chppz"] Nov 29 04:43:01 crc kubenswrapper[4799]: I1129 04:43:01.562985 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-69cd86ff6b-chppz" event={"ID":"55f389a3-e952-4cfb-8aec-4ae3aa563101","Type":"ContainerStarted","Data":"364d8b7abf4c290947774e6f083f834fa6dc0e312aa36623321f0ef2e7dcb76a"} Nov 29 04:43:01 crc kubenswrapper[4799]: I1129 04:43:01.563946 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-69cd86ff6b-chppz" event={"ID":"55f389a3-e952-4cfb-8aec-4ae3aa563101","Type":"ContainerStarted","Data":"c3f3d28351b509732dbcd1f82b674e373520b112c067be7da85c4dd1232b2720"} Nov 29 04:43:01 crc kubenswrapper[4799]: I1129 04:43:01.563991 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-69cd86ff6b-chppz" Nov 29 04:43:01 crc kubenswrapper[4799]: I1129 04:43:01.567020 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7b6db79c5b-4bndf" Nov 29 04:43:01 crc kubenswrapper[4799]: I1129 04:43:01.619755 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-69cd86ff6b-chppz" podStartSLOduration=1.619673656 podStartE2EDuration="1.619673656s" podCreationTimestamp="2025-11-29 04:43:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:43:01.589087898 +0000 UTC m=+257.232018328" watchObservedRunningTime="2025-11-29 04:43:01.619673656 +0000 UTC m=+257.262604086" Nov 29 04:43:01 crc kubenswrapper[4799]: I1129 04:43:01.641296 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-594766d55d-g5c8f"] Nov 29 04:43:01 crc kubenswrapper[4799]: I1129 04:43:01.642989 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-594766d55d-g5c8f" Nov 29 04:43:01 crc kubenswrapper[4799]: I1129 04:43:01.646751 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 29 04:43:01 crc kubenswrapper[4799]: I1129 04:43:01.647060 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 29 04:43:01 crc kubenswrapper[4799]: I1129 04:43:01.647369 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 29 04:43:01 crc kubenswrapper[4799]: I1129 04:43:01.647427 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 29 04:43:01 crc kubenswrapper[4799]: I1129 04:43:01.651488 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7b6db79c5b-4bndf"] Nov 29 04:43:01 crc kubenswrapper[4799]: I1129 04:43:01.656161 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-7b6db79c5b-4bndf"] Nov 29 04:43:01 crc kubenswrapper[4799]: I1129 04:43:01.660992 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 29 04:43:01 crc kubenswrapper[4799]: I1129 04:43:01.662287 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 29 04:43:01 crc kubenswrapper[4799]: I1129 04:43:01.663447 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-594766d55d-g5c8f"] Nov 29 04:43:01 crc kubenswrapper[4799]: I1129 04:43:01.664370 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 29 04:43:01 crc kubenswrapper[4799]: I1129 04:43:01.752034 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0fa71a14-00a9-4874-b41c-03254faa648c-client-ca\") pod \"controller-manager-594766d55d-g5c8f\" (UID: \"0fa71a14-00a9-4874-b41c-03254faa648c\") " pod="openshift-controller-manager/controller-manager-594766d55d-g5c8f" Nov 29 04:43:01 crc kubenswrapper[4799]: I1129 04:43:01.752087 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0fa71a14-00a9-4874-b41c-03254faa648c-config\") pod \"controller-manager-594766d55d-g5c8f\" (UID: \"0fa71a14-00a9-4874-b41c-03254faa648c\") " pod="openshift-controller-manager/controller-manager-594766d55d-g5c8f" Nov 29 04:43:01 crc kubenswrapper[4799]: I1129 04:43:01.752123 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0fa71a14-00a9-4874-b41c-03254faa648c-serving-cert\") pod \"controller-manager-594766d55d-g5c8f\" (UID: \"0fa71a14-00a9-4874-b41c-03254faa648c\") " pod="openshift-controller-manager/controller-manager-594766d55d-g5c8f" Nov 29 04:43:01 crc kubenswrapper[4799]: I1129 04:43:01.752153 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0fa71a14-00a9-4874-b41c-03254faa648c-proxy-ca-bundles\") pod \"controller-manager-594766d55d-g5c8f\" (UID: \"0fa71a14-00a9-4874-b41c-03254faa648c\") " pod="openshift-controller-manager/controller-manager-594766d55d-g5c8f" Nov 29 04:43:01 crc kubenswrapper[4799]: I1129 04:43:01.752183 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2pkn\" (UniqueName: \"kubernetes.io/projected/0fa71a14-00a9-4874-b41c-03254faa648c-kube-api-access-j2pkn\") pod \"controller-manager-594766d55d-g5c8f\" (UID: \"0fa71a14-00a9-4874-b41c-03254faa648c\") " pod="openshift-controller-manager/controller-manager-594766d55d-g5c8f" Nov 29 04:43:01 crc kubenswrapper[4799]: I1129 04:43:01.853689 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0fa71a14-00a9-4874-b41c-03254faa648c-client-ca\") pod \"controller-manager-594766d55d-g5c8f\" (UID: \"0fa71a14-00a9-4874-b41c-03254faa648c\") " pod="openshift-controller-manager/controller-manager-594766d55d-g5c8f" Nov 29 04:43:01 crc kubenswrapper[4799]: I1129 04:43:01.853759 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0fa71a14-00a9-4874-b41c-03254faa648c-config\") pod \"controller-manager-594766d55d-g5c8f\" (UID: \"0fa71a14-00a9-4874-b41c-03254faa648c\") " pod="openshift-controller-manager/controller-manager-594766d55d-g5c8f" Nov 29 04:43:01 crc kubenswrapper[4799]: I1129 04:43:01.853889 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0fa71a14-00a9-4874-b41c-03254faa648c-serving-cert\") pod \"controller-manager-594766d55d-g5c8f\" (UID: \"0fa71a14-00a9-4874-b41c-03254faa648c\") " pod="openshift-controller-manager/controller-manager-594766d55d-g5c8f" Nov 29 04:43:01 crc kubenswrapper[4799]: I1129 04:43:01.853942 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0fa71a14-00a9-4874-b41c-03254faa648c-proxy-ca-bundles\") pod \"controller-manager-594766d55d-g5c8f\" (UID: \"0fa71a14-00a9-4874-b41c-03254faa648c\") " pod="openshift-controller-manager/controller-manager-594766d55d-g5c8f" Nov 29 04:43:01 crc kubenswrapper[4799]: I1129 04:43:01.854081 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2pkn\" (UniqueName: \"kubernetes.io/projected/0fa71a14-00a9-4874-b41c-03254faa648c-kube-api-access-j2pkn\") pod \"controller-manager-594766d55d-g5c8f\" (UID: \"0fa71a14-00a9-4874-b41c-03254faa648c\") " pod="openshift-controller-manager/controller-manager-594766d55d-g5c8f" Nov 29 04:43:01 crc kubenswrapper[4799]: I1129 04:43:01.855624 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0fa71a14-00a9-4874-b41c-03254faa648c-client-ca\") pod \"controller-manager-594766d55d-g5c8f\" (UID: \"0fa71a14-00a9-4874-b41c-03254faa648c\") " pod="openshift-controller-manager/controller-manager-594766d55d-g5c8f" Nov 29 04:43:01 crc kubenswrapper[4799]: I1129 04:43:01.856221 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0fa71a14-00a9-4874-b41c-03254faa648c-proxy-ca-bundles\") pod \"controller-manager-594766d55d-g5c8f\" (UID: \"0fa71a14-00a9-4874-b41c-03254faa648c\") " pod="openshift-controller-manager/controller-manager-594766d55d-g5c8f" Nov 29 04:43:01 crc kubenswrapper[4799]: I1129 04:43:01.856770 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0fa71a14-00a9-4874-b41c-03254faa648c-config\") pod \"controller-manager-594766d55d-g5c8f\" (UID: \"0fa71a14-00a9-4874-b41c-03254faa648c\") " pod="openshift-controller-manager/controller-manager-594766d55d-g5c8f" Nov 29 04:43:01 crc kubenswrapper[4799]: I1129 04:43:01.868278 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0fa71a14-00a9-4874-b41c-03254faa648c-serving-cert\") pod \"controller-manager-594766d55d-g5c8f\" (UID: \"0fa71a14-00a9-4874-b41c-03254faa648c\") " pod="openshift-controller-manager/controller-manager-594766d55d-g5c8f" Nov 29 04:43:01 crc kubenswrapper[4799]: I1129 04:43:01.872783 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2pkn\" (UniqueName: \"kubernetes.io/projected/0fa71a14-00a9-4874-b41c-03254faa648c-kube-api-access-j2pkn\") pod \"controller-manager-594766d55d-g5c8f\" (UID: \"0fa71a14-00a9-4874-b41c-03254faa648c\") " pod="openshift-controller-manager/controller-manager-594766d55d-g5c8f" Nov 29 04:43:01 crc kubenswrapper[4799]: I1129 04:43:01.965536 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-594766d55d-g5c8f" Nov 29 04:43:01 crc kubenswrapper[4799]: I1129 04:43:01.985918 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-69cd86ff6b-chppz" Nov 29 04:43:02 crc kubenswrapper[4799]: I1129 04:43:02.310520 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-594766d55d-g5c8f"] Nov 29 04:43:02 crc kubenswrapper[4799]: I1129 04:43:02.575278 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-594766d55d-g5c8f" event={"ID":"0fa71a14-00a9-4874-b41c-03254faa648c","Type":"ContainerStarted","Data":"7f1071d7c75025575076f5abfd35131323ab772325d065d8970478b1d71c5cca"} Nov 29 04:43:02 crc kubenswrapper[4799]: I1129 04:43:02.575819 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-594766d55d-g5c8f" event={"ID":"0fa71a14-00a9-4874-b41c-03254faa648c","Type":"ContainerStarted","Data":"942e805db5f5743985fcb97fcd2f2c17ab4b4e5f739c18c3332ed7e93014657f"} Nov 29 04:43:02 crc kubenswrapper[4799]: I1129 04:43:02.592816 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-594766d55d-g5c8f" podStartSLOduration=2.592763248 podStartE2EDuration="2.592763248s" podCreationTimestamp="2025-11-29 04:43:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:43:02.590350045 +0000 UTC m=+258.233280485" watchObservedRunningTime="2025-11-29 04:43:02.592763248 +0000 UTC m=+258.235693658" Nov 29 04:43:02 crc kubenswrapper[4799]: I1129 04:43:02.665760 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d10019fa-babc-4de4-be1b-df08d7247bc8" path="/var/lib/kubelet/pods/d10019fa-babc-4de4-be1b-df08d7247bc8/volumes" Nov 29 04:43:03 crc kubenswrapper[4799]: I1129 04:43:03.016955 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-446gh"] Nov 29 04:43:03 crc kubenswrapper[4799]: I1129 04:43:03.018115 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-446gh" Nov 29 04:43:03 crc kubenswrapper[4799]: I1129 04:43:03.020576 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 29 04:43:03 crc kubenswrapper[4799]: I1129 04:43:03.036591 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-446gh"] Nov 29 04:43:03 crc kubenswrapper[4799]: I1129 04:43:03.173280 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5133ffab-cfbb-4b1d-a161-d88f42ef37ce-utilities\") pod \"community-operators-446gh\" (UID: \"5133ffab-cfbb-4b1d-a161-d88f42ef37ce\") " pod="openshift-marketplace/community-operators-446gh" Nov 29 04:43:03 crc kubenswrapper[4799]: I1129 04:43:03.173360 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5133ffab-cfbb-4b1d-a161-d88f42ef37ce-catalog-content\") pod \"community-operators-446gh\" (UID: \"5133ffab-cfbb-4b1d-a161-d88f42ef37ce\") " pod="openshift-marketplace/community-operators-446gh" Nov 29 04:43:03 crc kubenswrapper[4799]: I1129 04:43:03.173430 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdxb8\" (UniqueName: \"kubernetes.io/projected/5133ffab-cfbb-4b1d-a161-d88f42ef37ce-kube-api-access-zdxb8\") pod \"community-operators-446gh\" (UID: \"5133ffab-cfbb-4b1d-a161-d88f42ef37ce\") " pod="openshift-marketplace/community-operators-446gh" Nov 29 04:43:03 crc kubenswrapper[4799]: I1129 04:43:03.217566 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-pmw7d"] Nov 29 04:43:03 crc kubenswrapper[4799]: I1129 04:43:03.221774 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pmw7d" Nov 29 04:43:03 crc kubenswrapper[4799]: I1129 04:43:03.224713 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 29 04:43:03 crc kubenswrapper[4799]: I1129 04:43:03.240573 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pmw7d"] Nov 29 04:43:03 crc kubenswrapper[4799]: I1129 04:43:03.274532 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5133ffab-cfbb-4b1d-a161-d88f42ef37ce-utilities\") pod \"community-operators-446gh\" (UID: \"5133ffab-cfbb-4b1d-a161-d88f42ef37ce\") " pod="openshift-marketplace/community-operators-446gh" Nov 29 04:43:03 crc kubenswrapper[4799]: I1129 04:43:03.274585 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5133ffab-cfbb-4b1d-a161-d88f42ef37ce-catalog-content\") pod \"community-operators-446gh\" (UID: \"5133ffab-cfbb-4b1d-a161-d88f42ef37ce\") " pod="openshift-marketplace/community-operators-446gh" Nov 29 04:43:03 crc kubenswrapper[4799]: I1129 04:43:03.274619 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdxb8\" (UniqueName: \"kubernetes.io/projected/5133ffab-cfbb-4b1d-a161-d88f42ef37ce-kube-api-access-zdxb8\") pod \"community-operators-446gh\" (UID: \"5133ffab-cfbb-4b1d-a161-d88f42ef37ce\") " pod="openshift-marketplace/community-operators-446gh" Nov 29 04:43:03 crc kubenswrapper[4799]: I1129 04:43:03.275607 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5133ffab-cfbb-4b1d-a161-d88f42ef37ce-utilities\") pod \"community-operators-446gh\" (UID: \"5133ffab-cfbb-4b1d-a161-d88f42ef37ce\") " pod="openshift-marketplace/community-operators-446gh" Nov 29 04:43:03 crc kubenswrapper[4799]: I1129 04:43:03.276238 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5133ffab-cfbb-4b1d-a161-d88f42ef37ce-catalog-content\") pod \"community-operators-446gh\" (UID: \"5133ffab-cfbb-4b1d-a161-d88f42ef37ce\") " pod="openshift-marketplace/community-operators-446gh" Nov 29 04:43:03 crc kubenswrapper[4799]: I1129 04:43:03.303985 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdxb8\" (UniqueName: \"kubernetes.io/projected/5133ffab-cfbb-4b1d-a161-d88f42ef37ce-kube-api-access-zdxb8\") pod \"community-operators-446gh\" (UID: \"5133ffab-cfbb-4b1d-a161-d88f42ef37ce\") " pod="openshift-marketplace/community-operators-446gh" Nov 29 04:43:03 crc kubenswrapper[4799]: I1129 04:43:03.337528 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-446gh" Nov 29 04:43:03 crc kubenswrapper[4799]: I1129 04:43:03.376786 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfabf29e-ad4c-4ded-b754-5b02a25ff34b-utilities\") pod \"certified-operators-pmw7d\" (UID: \"cfabf29e-ad4c-4ded-b754-5b02a25ff34b\") " pod="openshift-marketplace/certified-operators-pmw7d" Nov 29 04:43:03 crc kubenswrapper[4799]: I1129 04:43:03.377212 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p88sn\" (UniqueName: \"kubernetes.io/projected/cfabf29e-ad4c-4ded-b754-5b02a25ff34b-kube-api-access-p88sn\") pod \"certified-operators-pmw7d\" (UID: \"cfabf29e-ad4c-4ded-b754-5b02a25ff34b\") " pod="openshift-marketplace/certified-operators-pmw7d" Nov 29 04:43:03 crc kubenswrapper[4799]: I1129 04:43:03.377720 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfabf29e-ad4c-4ded-b754-5b02a25ff34b-catalog-content\") pod \"certified-operators-pmw7d\" (UID: \"cfabf29e-ad4c-4ded-b754-5b02a25ff34b\") " pod="openshift-marketplace/certified-operators-pmw7d" Nov 29 04:43:03 crc kubenswrapper[4799]: I1129 04:43:03.479587 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p88sn\" (UniqueName: \"kubernetes.io/projected/cfabf29e-ad4c-4ded-b754-5b02a25ff34b-kube-api-access-p88sn\") pod \"certified-operators-pmw7d\" (UID: \"cfabf29e-ad4c-4ded-b754-5b02a25ff34b\") " pod="openshift-marketplace/certified-operators-pmw7d" Nov 29 04:43:03 crc kubenswrapper[4799]: I1129 04:43:03.480123 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfabf29e-ad4c-4ded-b754-5b02a25ff34b-catalog-content\") pod \"certified-operators-pmw7d\" (UID: \"cfabf29e-ad4c-4ded-b754-5b02a25ff34b\") " pod="openshift-marketplace/certified-operators-pmw7d" Nov 29 04:43:03 crc kubenswrapper[4799]: I1129 04:43:03.480214 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfabf29e-ad4c-4ded-b754-5b02a25ff34b-utilities\") pod \"certified-operators-pmw7d\" (UID: \"cfabf29e-ad4c-4ded-b754-5b02a25ff34b\") " pod="openshift-marketplace/certified-operators-pmw7d" Nov 29 04:43:03 crc kubenswrapper[4799]: I1129 04:43:03.480837 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfabf29e-ad4c-4ded-b754-5b02a25ff34b-utilities\") pod \"certified-operators-pmw7d\" (UID: \"cfabf29e-ad4c-4ded-b754-5b02a25ff34b\") " pod="openshift-marketplace/certified-operators-pmw7d" Nov 29 04:43:03 crc kubenswrapper[4799]: I1129 04:43:03.480904 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfabf29e-ad4c-4ded-b754-5b02a25ff34b-catalog-content\") pod \"certified-operators-pmw7d\" (UID: \"cfabf29e-ad4c-4ded-b754-5b02a25ff34b\") " pod="openshift-marketplace/certified-operators-pmw7d" Nov 29 04:43:03 crc kubenswrapper[4799]: I1129 04:43:03.502258 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p88sn\" (UniqueName: \"kubernetes.io/projected/cfabf29e-ad4c-4ded-b754-5b02a25ff34b-kube-api-access-p88sn\") pod \"certified-operators-pmw7d\" (UID: \"cfabf29e-ad4c-4ded-b754-5b02a25ff34b\") " pod="openshift-marketplace/certified-operators-pmw7d" Nov 29 04:43:03 crc kubenswrapper[4799]: I1129 04:43:03.551012 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-446gh"] Nov 29 04:43:03 crc kubenswrapper[4799]: I1129 04:43:03.590916 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-446gh" event={"ID":"5133ffab-cfbb-4b1d-a161-d88f42ef37ce","Type":"ContainerStarted","Data":"773f0f87567ce61a58c274605115e369dbcb8919ed32e615e3d43f97b6543330"} Nov 29 04:43:03 crc kubenswrapper[4799]: I1129 04:43:03.592898 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-594766d55d-g5c8f" Nov 29 04:43:03 crc kubenswrapper[4799]: I1129 04:43:03.600886 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-594766d55d-g5c8f" Nov 29 04:43:03 crc kubenswrapper[4799]: I1129 04:43:03.625246 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pmw7d" Nov 29 04:43:03 crc kubenswrapper[4799]: I1129 04:43:03.837184 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pmw7d"] Nov 29 04:43:03 crc kubenswrapper[4799]: W1129 04:43:03.846885 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcfabf29e_ad4c_4ded_b754_5b02a25ff34b.slice/crio-06eb308127fe0963b45186c4dc019d03a7cae30481dbeeb42dce0f910926e237 WatchSource:0}: Error finding container 06eb308127fe0963b45186c4dc019d03a7cae30481dbeeb42dce0f910926e237: Status 404 returned error can't find the container with id 06eb308127fe0963b45186c4dc019d03a7cae30481dbeeb42dce0f910926e237 Nov 29 04:43:04 crc kubenswrapper[4799]: I1129 04:43:04.598440 4799 generic.go:334] "Generic (PLEG): container finished" podID="cfabf29e-ad4c-4ded-b754-5b02a25ff34b" containerID="ab67b7125632d39bc654d311ede5e3a8d4c922ca4eb08302920183c821f5adf8" exitCode=0 Nov 29 04:43:04 crc kubenswrapper[4799]: I1129 04:43:04.598836 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pmw7d" event={"ID":"cfabf29e-ad4c-4ded-b754-5b02a25ff34b","Type":"ContainerDied","Data":"ab67b7125632d39bc654d311ede5e3a8d4c922ca4eb08302920183c821f5adf8"} Nov 29 04:43:04 crc kubenswrapper[4799]: I1129 04:43:04.598931 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pmw7d" event={"ID":"cfabf29e-ad4c-4ded-b754-5b02a25ff34b","Type":"ContainerStarted","Data":"06eb308127fe0963b45186c4dc019d03a7cae30481dbeeb42dce0f910926e237"} Nov 29 04:43:04 crc kubenswrapper[4799]: I1129 04:43:04.601360 4799 generic.go:334] "Generic (PLEG): container finished" podID="5133ffab-cfbb-4b1d-a161-d88f42ef37ce" containerID="0132be20cfe9fdf4df45dedb92807009b0008025be3bc156101b9e65e89a2063" exitCode=0 Nov 29 04:43:04 crc kubenswrapper[4799]: I1129 04:43:04.601491 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-446gh" event={"ID":"5133ffab-cfbb-4b1d-a161-d88f42ef37ce","Type":"ContainerDied","Data":"0132be20cfe9fdf4df45dedb92807009b0008025be3bc156101b9e65e89a2063"} Nov 29 04:43:05 crc kubenswrapper[4799]: I1129 04:43:05.043309 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 29 04:43:05 crc kubenswrapper[4799]: I1129 04:43:05.415935 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qwxcg"] Nov 29 04:43:05 crc kubenswrapper[4799]: I1129 04:43:05.417048 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qwxcg" Nov 29 04:43:05 crc kubenswrapper[4799]: I1129 04:43:05.419407 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 29 04:43:05 crc kubenswrapper[4799]: I1129 04:43:05.426667 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qwxcg"] Nov 29 04:43:05 crc kubenswrapper[4799]: I1129 04:43:05.619148 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbt7t\" (UniqueName: \"kubernetes.io/projected/6ee05926-5af3-4fac-9772-8e1484c4de17-kube-api-access-gbt7t\") pod \"redhat-marketplace-qwxcg\" (UID: \"6ee05926-5af3-4fac-9772-8e1484c4de17\") " pod="openshift-marketplace/redhat-marketplace-qwxcg" Nov 29 04:43:05 crc kubenswrapper[4799]: I1129 04:43:05.619304 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ee05926-5af3-4fac-9772-8e1484c4de17-catalog-content\") pod \"redhat-marketplace-qwxcg\" (UID: \"6ee05926-5af3-4fac-9772-8e1484c4de17\") " pod="openshift-marketplace/redhat-marketplace-qwxcg" Nov 29 04:43:05 crc kubenswrapper[4799]: I1129 04:43:05.619414 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ee05926-5af3-4fac-9772-8e1484c4de17-utilities\") pod \"redhat-marketplace-qwxcg\" (UID: \"6ee05926-5af3-4fac-9772-8e1484c4de17\") " pod="openshift-marketplace/redhat-marketplace-qwxcg" Nov 29 04:43:05 crc kubenswrapper[4799]: I1129 04:43:05.625890 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xngqr"] Nov 29 04:43:05 crc kubenswrapper[4799]: I1129 04:43:05.627093 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xngqr" Nov 29 04:43:05 crc kubenswrapper[4799]: I1129 04:43:05.634889 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xngqr"] Nov 29 04:43:05 crc kubenswrapper[4799]: I1129 04:43:05.635346 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 29 04:43:05 crc kubenswrapper[4799]: I1129 04:43:05.720336 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47f0c357-1fc0-4646-ae33-b4497ac972e7-catalog-content\") pod \"redhat-operators-xngqr\" (UID: \"47f0c357-1fc0-4646-ae33-b4497ac972e7\") " pod="openshift-marketplace/redhat-operators-xngqr" Nov 29 04:43:05 crc kubenswrapper[4799]: I1129 04:43:05.720416 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbt7t\" (UniqueName: \"kubernetes.io/projected/6ee05926-5af3-4fac-9772-8e1484c4de17-kube-api-access-gbt7t\") pod \"redhat-marketplace-qwxcg\" (UID: \"6ee05926-5af3-4fac-9772-8e1484c4de17\") " pod="openshift-marketplace/redhat-marketplace-qwxcg" Nov 29 04:43:05 crc kubenswrapper[4799]: I1129 04:43:05.720456 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ee05926-5af3-4fac-9772-8e1484c4de17-catalog-content\") pod \"redhat-marketplace-qwxcg\" (UID: \"6ee05926-5af3-4fac-9772-8e1484c4de17\") " pod="openshift-marketplace/redhat-marketplace-qwxcg" Nov 29 04:43:05 crc kubenswrapper[4799]: I1129 04:43:05.720490 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5h7vs\" (UniqueName: \"kubernetes.io/projected/47f0c357-1fc0-4646-ae33-b4497ac972e7-kube-api-access-5h7vs\") pod \"redhat-operators-xngqr\" (UID: \"47f0c357-1fc0-4646-ae33-b4497ac972e7\") " pod="openshift-marketplace/redhat-operators-xngqr" Nov 29 04:43:05 crc kubenswrapper[4799]: I1129 04:43:05.720524 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ee05926-5af3-4fac-9772-8e1484c4de17-utilities\") pod \"redhat-marketplace-qwxcg\" (UID: \"6ee05926-5af3-4fac-9772-8e1484c4de17\") " pod="openshift-marketplace/redhat-marketplace-qwxcg" Nov 29 04:43:05 crc kubenswrapper[4799]: I1129 04:43:05.720615 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47f0c357-1fc0-4646-ae33-b4497ac972e7-utilities\") pod \"redhat-operators-xngqr\" (UID: \"47f0c357-1fc0-4646-ae33-b4497ac972e7\") " pod="openshift-marketplace/redhat-operators-xngqr" Nov 29 04:43:05 crc kubenswrapper[4799]: I1129 04:43:05.726546 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ee05926-5af3-4fac-9772-8e1484c4de17-catalog-content\") pod \"redhat-marketplace-qwxcg\" (UID: \"6ee05926-5af3-4fac-9772-8e1484c4de17\") " pod="openshift-marketplace/redhat-marketplace-qwxcg" Nov 29 04:43:05 crc kubenswrapper[4799]: I1129 04:43:05.726544 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ee05926-5af3-4fac-9772-8e1484c4de17-utilities\") pod \"redhat-marketplace-qwxcg\" (UID: \"6ee05926-5af3-4fac-9772-8e1484c4de17\") " pod="openshift-marketplace/redhat-marketplace-qwxcg" Nov 29 04:43:05 crc kubenswrapper[4799]: I1129 04:43:05.756627 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbt7t\" (UniqueName: \"kubernetes.io/projected/6ee05926-5af3-4fac-9772-8e1484c4de17-kube-api-access-gbt7t\") pod \"redhat-marketplace-qwxcg\" (UID: \"6ee05926-5af3-4fac-9772-8e1484c4de17\") " pod="openshift-marketplace/redhat-marketplace-qwxcg" Nov 29 04:43:05 crc kubenswrapper[4799]: I1129 04:43:05.828106 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47f0c357-1fc0-4646-ae33-b4497ac972e7-catalog-content\") pod \"redhat-operators-xngqr\" (UID: \"47f0c357-1fc0-4646-ae33-b4497ac972e7\") " pod="openshift-marketplace/redhat-operators-xngqr" Nov 29 04:43:05 crc kubenswrapper[4799]: I1129 04:43:05.828976 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5h7vs\" (UniqueName: \"kubernetes.io/projected/47f0c357-1fc0-4646-ae33-b4497ac972e7-kube-api-access-5h7vs\") pod \"redhat-operators-xngqr\" (UID: \"47f0c357-1fc0-4646-ae33-b4497ac972e7\") " pod="openshift-marketplace/redhat-operators-xngqr" Nov 29 04:43:05 crc kubenswrapper[4799]: I1129 04:43:05.829095 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47f0c357-1fc0-4646-ae33-b4497ac972e7-utilities\") pod \"redhat-operators-xngqr\" (UID: \"47f0c357-1fc0-4646-ae33-b4497ac972e7\") " pod="openshift-marketplace/redhat-operators-xngqr" Nov 29 04:43:05 crc kubenswrapper[4799]: I1129 04:43:05.829001 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47f0c357-1fc0-4646-ae33-b4497ac972e7-catalog-content\") pod \"redhat-operators-xngqr\" (UID: \"47f0c357-1fc0-4646-ae33-b4497ac972e7\") " pod="openshift-marketplace/redhat-operators-xngqr" Nov 29 04:43:05 crc kubenswrapper[4799]: I1129 04:43:05.829384 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47f0c357-1fc0-4646-ae33-b4497ac972e7-utilities\") pod \"redhat-operators-xngqr\" (UID: \"47f0c357-1fc0-4646-ae33-b4497ac972e7\") " pod="openshift-marketplace/redhat-operators-xngqr" Nov 29 04:43:05 crc kubenswrapper[4799]: I1129 04:43:05.881135 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5h7vs\" (UniqueName: \"kubernetes.io/projected/47f0c357-1fc0-4646-ae33-b4497ac972e7-kube-api-access-5h7vs\") pod \"redhat-operators-xngqr\" (UID: \"47f0c357-1fc0-4646-ae33-b4497ac972e7\") " pod="openshift-marketplace/redhat-operators-xngqr" Nov 29 04:43:05 crc kubenswrapper[4799]: I1129 04:43:05.964842 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xngqr" Nov 29 04:43:06 crc kubenswrapper[4799]: I1129 04:43:06.034274 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qwxcg" Nov 29 04:43:06 crc kubenswrapper[4799]: I1129 04:43:06.405292 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xngqr"] Nov 29 04:43:06 crc kubenswrapper[4799]: I1129 04:43:06.529453 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qwxcg"] Nov 29 04:43:06 crc kubenswrapper[4799]: W1129 04:43:06.535215 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ee05926_5af3_4fac_9772_8e1484c4de17.slice/crio-e66c629d87f80afdc3c1a27b878c683b44d6799c8453441dbf7acb26a17dd38d WatchSource:0}: Error finding container e66c629d87f80afdc3c1a27b878c683b44d6799c8453441dbf7acb26a17dd38d: Status 404 returned error can't find the container with id e66c629d87f80afdc3c1a27b878c683b44d6799c8453441dbf7acb26a17dd38d Nov 29 04:43:06 crc kubenswrapper[4799]: I1129 04:43:06.630048 4799 generic.go:334] "Generic (PLEG): container finished" podID="5133ffab-cfbb-4b1d-a161-d88f42ef37ce" containerID="333991016bafcc9c6d8c9cafe90b47b91f9226151db9ecd3b2956128102352ec" exitCode=0 Nov 29 04:43:06 crc kubenswrapper[4799]: I1129 04:43:06.630133 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-446gh" event={"ID":"5133ffab-cfbb-4b1d-a161-d88f42ef37ce","Type":"ContainerDied","Data":"333991016bafcc9c6d8c9cafe90b47b91f9226151db9ecd3b2956128102352ec"} Nov 29 04:43:06 crc kubenswrapper[4799]: I1129 04:43:06.635640 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qwxcg" event={"ID":"6ee05926-5af3-4fac-9772-8e1484c4de17","Type":"ContainerStarted","Data":"e66c629d87f80afdc3c1a27b878c683b44d6799c8453441dbf7acb26a17dd38d"} Nov 29 04:43:06 crc kubenswrapper[4799]: I1129 04:43:06.637938 4799 generic.go:334] "Generic (PLEG): container finished" podID="cfabf29e-ad4c-4ded-b754-5b02a25ff34b" containerID="b2653c69c3c11a98814f903703f1c9a59c3272a4676d9e82c8d2088553b49b1b" exitCode=0 Nov 29 04:43:06 crc kubenswrapper[4799]: I1129 04:43:06.638018 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pmw7d" event={"ID":"cfabf29e-ad4c-4ded-b754-5b02a25ff34b","Type":"ContainerDied","Data":"b2653c69c3c11a98814f903703f1c9a59c3272a4676d9e82c8d2088553b49b1b"} Nov 29 04:43:06 crc kubenswrapper[4799]: I1129 04:43:06.640441 4799 generic.go:334] "Generic (PLEG): container finished" podID="47f0c357-1fc0-4646-ae33-b4497ac972e7" containerID="733025811e468c4edb50ea9c1e33209cdaeae570d1093530b96e7633300d734d" exitCode=0 Nov 29 04:43:06 crc kubenswrapper[4799]: I1129 04:43:06.640479 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xngqr" event={"ID":"47f0c357-1fc0-4646-ae33-b4497ac972e7","Type":"ContainerDied","Data":"733025811e468c4edb50ea9c1e33209cdaeae570d1093530b96e7633300d734d"} Nov 29 04:43:06 crc kubenswrapper[4799]: I1129 04:43:06.640520 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xngqr" event={"ID":"47f0c357-1fc0-4646-ae33-b4497ac972e7","Type":"ContainerStarted","Data":"c9efa54dd23fada8bee4a96e9bde1ea7eef0e678b035bc4f08728801a14f6be0"} Nov 29 04:43:07 crc kubenswrapper[4799]: I1129 04:43:07.685609 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-446gh" event={"ID":"5133ffab-cfbb-4b1d-a161-d88f42ef37ce","Type":"ContainerStarted","Data":"3a082d038a2524fac8bc987dd4406cce128a3bc2fcaad85adee60c938e979894"} Nov 29 04:43:07 crc kubenswrapper[4799]: I1129 04:43:07.688713 4799 generic.go:334] "Generic (PLEG): container finished" podID="6ee05926-5af3-4fac-9772-8e1484c4de17" containerID="aba42f8f1cc6972bc54371735991ad94ceba5b73de504b2112a2be739fc1b805" exitCode=0 Nov 29 04:43:07 crc kubenswrapper[4799]: I1129 04:43:07.688842 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qwxcg" event={"ID":"6ee05926-5af3-4fac-9772-8e1484c4de17","Type":"ContainerDied","Data":"aba42f8f1cc6972bc54371735991ad94ceba5b73de504b2112a2be739fc1b805"} Nov 29 04:43:07 crc kubenswrapper[4799]: I1129 04:43:07.695671 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pmw7d" event={"ID":"cfabf29e-ad4c-4ded-b754-5b02a25ff34b","Type":"ContainerStarted","Data":"8da837250b9293db221f661274170715e5ccccec164c2542d8e20f16102ef4c7"} Nov 29 04:43:07 crc kubenswrapper[4799]: I1129 04:43:07.698664 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xngqr" event={"ID":"47f0c357-1fc0-4646-ae33-b4497ac972e7","Type":"ContainerStarted","Data":"4885980c850437938d3e39caf94dafcf97753d3d96d4a89db8acd5c49f1b4351"} Nov 29 04:43:07 crc kubenswrapper[4799]: I1129 04:43:07.710545 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-446gh" podStartSLOduration=2.206393059 podStartE2EDuration="4.710523811s" podCreationTimestamp="2025-11-29 04:43:03 +0000 UTC" firstStartedPulling="2025-11-29 04:43:04.60286718 +0000 UTC m=+260.245797580" lastFinishedPulling="2025-11-29 04:43:07.106997892 +0000 UTC m=+262.749928332" observedRunningTime="2025-11-29 04:43:07.709618497 +0000 UTC m=+263.352548897" watchObservedRunningTime="2025-11-29 04:43:07.710523811 +0000 UTC m=+263.353454211" Nov 29 04:43:07 crc kubenswrapper[4799]: I1129 04:43:07.813131 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-pmw7d" podStartSLOduration=2.192369463 podStartE2EDuration="4.813103973s" podCreationTimestamp="2025-11-29 04:43:03 +0000 UTC" firstStartedPulling="2025-11-29 04:43:04.600927219 +0000 UTC m=+260.243857619" lastFinishedPulling="2025-11-29 04:43:07.221661729 +0000 UTC m=+262.864592129" observedRunningTime="2025-11-29 04:43:07.812165349 +0000 UTC m=+263.455095779" watchObservedRunningTime="2025-11-29 04:43:07.813103973 +0000 UTC m=+263.456034373" Nov 29 04:43:08 crc kubenswrapper[4799]: I1129 04:43:08.706974 4799 generic.go:334] "Generic (PLEG): container finished" podID="47f0c357-1fc0-4646-ae33-b4497ac972e7" containerID="4885980c850437938d3e39caf94dafcf97753d3d96d4a89db8acd5c49f1b4351" exitCode=0 Nov 29 04:43:08 crc kubenswrapper[4799]: I1129 04:43:08.707085 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xngqr" event={"ID":"47f0c357-1fc0-4646-ae33-b4497ac972e7","Type":"ContainerDied","Data":"4885980c850437938d3e39caf94dafcf97753d3d96d4a89db8acd5c49f1b4351"} Nov 29 04:43:08 crc kubenswrapper[4799]: I1129 04:43:08.710913 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qwxcg" event={"ID":"6ee05926-5af3-4fac-9772-8e1484c4de17","Type":"ContainerStarted","Data":"5d9c48eb02d82560105657e4d3e81496c0d0fde365ec456224c1c3d75cff2ad3"} Nov 29 04:43:09 crc kubenswrapper[4799]: I1129 04:43:09.717326 4799 generic.go:334] "Generic (PLEG): container finished" podID="6ee05926-5af3-4fac-9772-8e1484c4de17" containerID="5d9c48eb02d82560105657e4d3e81496c0d0fde365ec456224c1c3d75cff2ad3" exitCode=0 Nov 29 04:43:09 crc kubenswrapper[4799]: I1129 04:43:09.717401 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qwxcg" event={"ID":"6ee05926-5af3-4fac-9772-8e1484c4de17","Type":"ContainerDied","Data":"5d9c48eb02d82560105657e4d3e81496c0d0fde365ec456224c1c3d75cff2ad3"} Nov 29 04:43:09 crc kubenswrapper[4799]: I1129 04:43:09.720876 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xngqr" event={"ID":"47f0c357-1fc0-4646-ae33-b4497ac972e7","Type":"ContainerStarted","Data":"473952a11d3bb865828ca8169e9afd7352a0b20218adb62675733eb103660030"} Nov 29 04:43:09 crc kubenswrapper[4799]: I1129 04:43:09.765904 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xngqr" podStartSLOduration=2.199599365 podStartE2EDuration="4.765882191s" podCreationTimestamp="2025-11-29 04:43:05 +0000 UTC" firstStartedPulling="2025-11-29 04:43:06.641734052 +0000 UTC m=+262.284664452" lastFinishedPulling="2025-11-29 04:43:09.208016878 +0000 UTC m=+264.850947278" observedRunningTime="2025-11-29 04:43:09.76046411 +0000 UTC m=+265.403394510" watchObservedRunningTime="2025-11-29 04:43:09.765882191 +0000 UTC m=+265.408812591" Nov 29 04:43:10 crc kubenswrapper[4799]: I1129 04:43:10.730640 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qwxcg" event={"ID":"6ee05926-5af3-4fac-9772-8e1484c4de17","Type":"ContainerStarted","Data":"c96d3f870ae308729ff7f809f8af810f41ab755f1c7c3816fec899fad7efa29e"} Nov 29 04:43:10 crc kubenswrapper[4799]: I1129 04:43:10.757145 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qwxcg" podStartSLOduration=3.315752203 podStartE2EDuration="5.757121255s" podCreationTimestamp="2025-11-29 04:43:05 +0000 UTC" firstStartedPulling="2025-11-29 04:43:07.690541581 +0000 UTC m=+263.333471981" lastFinishedPulling="2025-11-29 04:43:10.131910613 +0000 UTC m=+265.774841033" observedRunningTime="2025-11-29 04:43:10.75539659 +0000 UTC m=+266.398326990" watchObservedRunningTime="2025-11-29 04:43:10.757121255 +0000 UTC m=+266.400051655" Nov 29 04:43:13 crc kubenswrapper[4799]: I1129 04:43:13.337776 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-446gh" Nov 29 04:43:13 crc kubenswrapper[4799]: I1129 04:43:13.338908 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-446gh" Nov 29 04:43:13 crc kubenswrapper[4799]: I1129 04:43:13.398954 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-446gh" Nov 29 04:43:13 crc kubenswrapper[4799]: I1129 04:43:13.626286 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-pmw7d" Nov 29 04:43:13 crc kubenswrapper[4799]: I1129 04:43:13.626373 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-pmw7d" Nov 29 04:43:13 crc kubenswrapper[4799]: I1129 04:43:13.685444 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-pmw7d" Nov 29 04:43:13 crc kubenswrapper[4799]: I1129 04:43:13.784454 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-pmw7d" Nov 29 04:43:13 crc kubenswrapper[4799]: I1129 04:43:13.795646 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-446gh" Nov 29 04:43:15 crc kubenswrapper[4799]: I1129 04:43:15.965570 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xngqr" Nov 29 04:43:15 crc kubenswrapper[4799]: I1129 04:43:15.966032 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xngqr" Nov 29 04:43:16 crc kubenswrapper[4799]: I1129 04:43:16.021991 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xngqr" Nov 29 04:43:16 crc kubenswrapper[4799]: I1129 04:43:16.037358 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qwxcg" Nov 29 04:43:16 crc kubenswrapper[4799]: I1129 04:43:16.037436 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qwxcg" Nov 29 04:43:16 crc kubenswrapper[4799]: I1129 04:43:16.091058 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qwxcg" Nov 29 04:43:16 crc kubenswrapper[4799]: I1129 04:43:16.800131 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qwxcg" Nov 29 04:43:16 crc kubenswrapper[4799]: I1129 04:43:16.801541 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xngqr" Nov 29 04:43:19 crc kubenswrapper[4799]: I1129 04:43:19.776497 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-594766d55d-g5c8f"] Nov 29 04:43:19 crc kubenswrapper[4799]: I1129 04:43:19.777004 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-594766d55d-g5c8f" podUID="0fa71a14-00a9-4874-b41c-03254faa648c" containerName="controller-manager" containerID="cri-o://7f1071d7c75025575076f5abfd35131323ab772325d065d8970478b1d71c5cca" gracePeriod=30 Nov 29 04:43:21 crc kubenswrapper[4799]: I1129 04:43:21.966994 4799 patch_prober.go:28] interesting pod/controller-manager-594766d55d-g5c8f container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.61:8443/healthz\": dial tcp 10.217.0.61:8443: connect: connection refused" start-of-body= Nov 29 04:43:21 crc kubenswrapper[4799]: I1129 04:43:21.967437 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-594766d55d-g5c8f" podUID="0fa71a14-00a9-4874-b41c-03254faa648c" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.61:8443/healthz\": dial tcp 10.217.0.61:8443: connect: connection refused" Nov 29 04:43:23 crc kubenswrapper[4799]: I1129 04:43:23.802356 4799 generic.go:334] "Generic (PLEG): container finished" podID="0fa71a14-00a9-4874-b41c-03254faa648c" containerID="7f1071d7c75025575076f5abfd35131323ab772325d065d8970478b1d71c5cca" exitCode=0 Nov 29 04:43:23 crc kubenswrapper[4799]: I1129 04:43:23.802432 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-594766d55d-g5c8f" event={"ID":"0fa71a14-00a9-4874-b41c-03254faa648c","Type":"ContainerDied","Data":"7f1071d7c75025575076f5abfd35131323ab772325d065d8970478b1d71c5cca"} Nov 29 04:43:24 crc kubenswrapper[4799]: I1129 04:43:24.062360 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-594766d55d-g5c8f" Nov 29 04:43:24 crc kubenswrapper[4799]: I1129 04:43:24.098631 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7b6db79c5b-mdbvh"] Nov 29 04:43:24 crc kubenswrapper[4799]: E1129 04:43:24.098938 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fa71a14-00a9-4874-b41c-03254faa648c" containerName="controller-manager" Nov 29 04:43:24 crc kubenswrapper[4799]: I1129 04:43:24.098956 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fa71a14-00a9-4874-b41c-03254faa648c" containerName="controller-manager" Nov 29 04:43:24 crc kubenswrapper[4799]: I1129 04:43:24.099124 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fa71a14-00a9-4874-b41c-03254faa648c" containerName="controller-manager" Nov 29 04:43:24 crc kubenswrapper[4799]: I1129 04:43:24.099699 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7b6db79c5b-mdbvh" Nov 29 04:43:24 crc kubenswrapper[4799]: I1129 04:43:24.117223 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7b6db79c5b-mdbvh"] Nov 29 04:43:24 crc kubenswrapper[4799]: I1129 04:43:24.141546 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0fa71a14-00a9-4874-b41c-03254faa648c-client-ca\") pod \"0fa71a14-00a9-4874-b41c-03254faa648c\" (UID: \"0fa71a14-00a9-4874-b41c-03254faa648c\") " Nov 29 04:43:24 crc kubenswrapper[4799]: I1129 04:43:24.141625 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0fa71a14-00a9-4874-b41c-03254faa648c-proxy-ca-bundles\") pod \"0fa71a14-00a9-4874-b41c-03254faa648c\" (UID: \"0fa71a14-00a9-4874-b41c-03254faa648c\") " Nov 29 04:43:24 crc kubenswrapper[4799]: I1129 04:43:24.141677 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0fa71a14-00a9-4874-b41c-03254faa648c-serving-cert\") pod \"0fa71a14-00a9-4874-b41c-03254faa648c\" (UID: \"0fa71a14-00a9-4874-b41c-03254faa648c\") " Nov 29 04:43:24 crc kubenswrapper[4799]: I1129 04:43:24.141718 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2pkn\" (UniqueName: \"kubernetes.io/projected/0fa71a14-00a9-4874-b41c-03254faa648c-kube-api-access-j2pkn\") pod \"0fa71a14-00a9-4874-b41c-03254faa648c\" (UID: \"0fa71a14-00a9-4874-b41c-03254faa648c\") " Nov 29 04:43:24 crc kubenswrapper[4799]: I1129 04:43:24.141823 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0fa71a14-00a9-4874-b41c-03254faa648c-config\") pod \"0fa71a14-00a9-4874-b41c-03254faa648c\" (UID: \"0fa71a14-00a9-4874-b41c-03254faa648c\") " Nov 29 04:43:24 crc kubenswrapper[4799]: I1129 04:43:24.142069 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/364f22df-4531-4a37-8657-b85af7e58eaa-config\") pod \"controller-manager-7b6db79c5b-mdbvh\" (UID: \"364f22df-4531-4a37-8657-b85af7e58eaa\") " pod="openshift-controller-manager/controller-manager-7b6db79c5b-mdbvh" Nov 29 04:43:24 crc kubenswrapper[4799]: I1129 04:43:24.142113 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/364f22df-4531-4a37-8657-b85af7e58eaa-proxy-ca-bundles\") pod \"controller-manager-7b6db79c5b-mdbvh\" (UID: \"364f22df-4531-4a37-8657-b85af7e58eaa\") " pod="openshift-controller-manager/controller-manager-7b6db79c5b-mdbvh" Nov 29 04:43:24 crc kubenswrapper[4799]: I1129 04:43:24.142187 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9db7p\" (UniqueName: \"kubernetes.io/projected/364f22df-4531-4a37-8657-b85af7e58eaa-kube-api-access-9db7p\") pod \"controller-manager-7b6db79c5b-mdbvh\" (UID: \"364f22df-4531-4a37-8657-b85af7e58eaa\") " pod="openshift-controller-manager/controller-manager-7b6db79c5b-mdbvh" Nov 29 04:43:24 crc kubenswrapper[4799]: I1129 04:43:24.142259 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/364f22df-4531-4a37-8657-b85af7e58eaa-serving-cert\") pod \"controller-manager-7b6db79c5b-mdbvh\" (UID: \"364f22df-4531-4a37-8657-b85af7e58eaa\") " pod="openshift-controller-manager/controller-manager-7b6db79c5b-mdbvh" Nov 29 04:43:24 crc kubenswrapper[4799]: I1129 04:43:24.142288 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/364f22df-4531-4a37-8657-b85af7e58eaa-client-ca\") pod \"controller-manager-7b6db79c5b-mdbvh\" (UID: \"364f22df-4531-4a37-8657-b85af7e58eaa\") " pod="openshift-controller-manager/controller-manager-7b6db79c5b-mdbvh" Nov 29 04:43:24 crc kubenswrapper[4799]: I1129 04:43:24.142682 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0fa71a14-00a9-4874-b41c-03254faa648c-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "0fa71a14-00a9-4874-b41c-03254faa648c" (UID: "0fa71a14-00a9-4874-b41c-03254faa648c"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:43:24 crc kubenswrapper[4799]: I1129 04:43:24.142717 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0fa71a14-00a9-4874-b41c-03254faa648c-client-ca" (OuterVolumeSpecName: "client-ca") pod "0fa71a14-00a9-4874-b41c-03254faa648c" (UID: "0fa71a14-00a9-4874-b41c-03254faa648c"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:43:24 crc kubenswrapper[4799]: I1129 04:43:24.142731 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0fa71a14-00a9-4874-b41c-03254faa648c-config" (OuterVolumeSpecName: "config") pod "0fa71a14-00a9-4874-b41c-03254faa648c" (UID: "0fa71a14-00a9-4874-b41c-03254faa648c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:43:24 crc kubenswrapper[4799]: I1129 04:43:24.146853 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fa71a14-00a9-4874-b41c-03254faa648c-kube-api-access-j2pkn" (OuterVolumeSpecName: "kube-api-access-j2pkn") pod "0fa71a14-00a9-4874-b41c-03254faa648c" (UID: "0fa71a14-00a9-4874-b41c-03254faa648c"). InnerVolumeSpecName "kube-api-access-j2pkn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:43:24 crc kubenswrapper[4799]: I1129 04:43:24.146874 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fa71a14-00a9-4874-b41c-03254faa648c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0fa71a14-00a9-4874-b41c-03254faa648c" (UID: "0fa71a14-00a9-4874-b41c-03254faa648c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:43:24 crc kubenswrapper[4799]: I1129 04:43:24.243554 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/364f22df-4531-4a37-8657-b85af7e58eaa-proxy-ca-bundles\") pod \"controller-manager-7b6db79c5b-mdbvh\" (UID: \"364f22df-4531-4a37-8657-b85af7e58eaa\") " pod="openshift-controller-manager/controller-manager-7b6db79c5b-mdbvh" Nov 29 04:43:24 crc kubenswrapper[4799]: I1129 04:43:24.243683 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9db7p\" (UniqueName: \"kubernetes.io/projected/364f22df-4531-4a37-8657-b85af7e58eaa-kube-api-access-9db7p\") pod \"controller-manager-7b6db79c5b-mdbvh\" (UID: \"364f22df-4531-4a37-8657-b85af7e58eaa\") " pod="openshift-controller-manager/controller-manager-7b6db79c5b-mdbvh" Nov 29 04:43:24 crc kubenswrapper[4799]: I1129 04:43:24.243758 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/364f22df-4531-4a37-8657-b85af7e58eaa-client-ca\") pod \"controller-manager-7b6db79c5b-mdbvh\" (UID: \"364f22df-4531-4a37-8657-b85af7e58eaa\") " pod="openshift-controller-manager/controller-manager-7b6db79c5b-mdbvh" Nov 29 04:43:24 crc kubenswrapper[4799]: I1129 04:43:24.243801 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/364f22df-4531-4a37-8657-b85af7e58eaa-serving-cert\") pod \"controller-manager-7b6db79c5b-mdbvh\" (UID: \"364f22df-4531-4a37-8657-b85af7e58eaa\") " pod="openshift-controller-manager/controller-manager-7b6db79c5b-mdbvh" Nov 29 04:43:24 crc kubenswrapper[4799]: I1129 04:43:24.243845 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/364f22df-4531-4a37-8657-b85af7e58eaa-config\") pod \"controller-manager-7b6db79c5b-mdbvh\" (UID: \"364f22df-4531-4a37-8657-b85af7e58eaa\") " pod="openshift-controller-manager/controller-manager-7b6db79c5b-mdbvh" Nov 29 04:43:24 crc kubenswrapper[4799]: I1129 04:43:24.243909 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0fa71a14-00a9-4874-b41c-03254faa648c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 04:43:24 crc kubenswrapper[4799]: I1129 04:43:24.243926 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2pkn\" (UniqueName: \"kubernetes.io/projected/0fa71a14-00a9-4874-b41c-03254faa648c-kube-api-access-j2pkn\") on node \"crc\" DevicePath \"\"" Nov 29 04:43:24 crc kubenswrapper[4799]: I1129 04:43:24.243940 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0fa71a14-00a9-4874-b41c-03254faa648c-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:43:24 crc kubenswrapper[4799]: I1129 04:43:24.243954 4799 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0fa71a14-00a9-4874-b41c-03254faa648c-client-ca\") on node \"crc\" DevicePath \"\"" Nov 29 04:43:24 crc kubenswrapper[4799]: I1129 04:43:24.243966 4799 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0fa71a14-00a9-4874-b41c-03254faa648c-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 29 04:43:24 crc kubenswrapper[4799]: I1129 04:43:24.245291 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/364f22df-4531-4a37-8657-b85af7e58eaa-client-ca\") pod \"controller-manager-7b6db79c5b-mdbvh\" (UID: \"364f22df-4531-4a37-8657-b85af7e58eaa\") " pod="openshift-controller-manager/controller-manager-7b6db79c5b-mdbvh" Nov 29 04:43:24 crc kubenswrapper[4799]: I1129 04:43:24.245383 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/364f22df-4531-4a37-8657-b85af7e58eaa-proxy-ca-bundles\") pod \"controller-manager-7b6db79c5b-mdbvh\" (UID: \"364f22df-4531-4a37-8657-b85af7e58eaa\") " pod="openshift-controller-manager/controller-manager-7b6db79c5b-mdbvh" Nov 29 04:43:24 crc kubenswrapper[4799]: I1129 04:43:24.245708 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/364f22df-4531-4a37-8657-b85af7e58eaa-config\") pod \"controller-manager-7b6db79c5b-mdbvh\" (UID: \"364f22df-4531-4a37-8657-b85af7e58eaa\") " pod="openshift-controller-manager/controller-manager-7b6db79c5b-mdbvh" Nov 29 04:43:24 crc kubenswrapper[4799]: I1129 04:43:24.249151 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/364f22df-4531-4a37-8657-b85af7e58eaa-serving-cert\") pod \"controller-manager-7b6db79c5b-mdbvh\" (UID: \"364f22df-4531-4a37-8657-b85af7e58eaa\") " pod="openshift-controller-manager/controller-manager-7b6db79c5b-mdbvh" Nov 29 04:43:24 crc kubenswrapper[4799]: I1129 04:43:24.265661 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9db7p\" (UniqueName: \"kubernetes.io/projected/364f22df-4531-4a37-8657-b85af7e58eaa-kube-api-access-9db7p\") pod \"controller-manager-7b6db79c5b-mdbvh\" (UID: \"364f22df-4531-4a37-8657-b85af7e58eaa\") " pod="openshift-controller-manager/controller-manager-7b6db79c5b-mdbvh" Nov 29 04:43:24 crc kubenswrapper[4799]: I1129 04:43:24.426075 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7b6db79c5b-mdbvh" Nov 29 04:43:24 crc kubenswrapper[4799]: I1129 04:43:24.820371 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-594766d55d-g5c8f" event={"ID":"0fa71a14-00a9-4874-b41c-03254faa648c","Type":"ContainerDied","Data":"942e805db5f5743985fcb97fcd2f2c17ab4b4e5f739c18c3332ed7e93014657f"} Nov 29 04:43:24 crc kubenswrapper[4799]: I1129 04:43:24.820450 4799 scope.go:117] "RemoveContainer" containerID="7f1071d7c75025575076f5abfd35131323ab772325d065d8970478b1d71c5cca" Nov 29 04:43:24 crc kubenswrapper[4799]: I1129 04:43:24.820533 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-594766d55d-g5c8f" Nov 29 04:43:24 crc kubenswrapper[4799]: I1129 04:43:24.840101 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-594766d55d-g5c8f"] Nov 29 04:43:24 crc kubenswrapper[4799]: I1129 04:43:24.844841 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-594766d55d-g5c8f"] Nov 29 04:43:24 crc kubenswrapper[4799]: W1129 04:43:24.855463 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod364f22df_4531_4a37_8657_b85af7e58eaa.slice/crio-5b6cbd4d865c3a9327d0918598bbe1df1c79bb5ba473092004a0e389eca4e4b5 WatchSource:0}: Error finding container 5b6cbd4d865c3a9327d0918598bbe1df1c79bb5ba473092004a0e389eca4e4b5: Status 404 returned error can't find the container with id 5b6cbd4d865c3a9327d0918598bbe1df1c79bb5ba473092004a0e389eca4e4b5 Nov 29 04:43:24 crc kubenswrapper[4799]: I1129 04:43:24.855573 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7b6db79c5b-mdbvh"] Nov 29 04:43:25 crc kubenswrapper[4799]: I1129 04:43:25.830813 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7b6db79c5b-mdbvh" event={"ID":"364f22df-4531-4a37-8657-b85af7e58eaa","Type":"ContainerStarted","Data":"b104532123fbbea4ec1c51a24ff7a93fc88e20057a60db31b73a0a7e32a7ec84"} Nov 29 04:43:25 crc kubenswrapper[4799]: I1129 04:43:25.831336 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7b6db79c5b-mdbvh" event={"ID":"364f22df-4531-4a37-8657-b85af7e58eaa","Type":"ContainerStarted","Data":"5b6cbd4d865c3a9327d0918598bbe1df1c79bb5ba473092004a0e389eca4e4b5"} Nov 29 04:43:25 crc kubenswrapper[4799]: I1129 04:43:25.831358 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7b6db79c5b-mdbvh" Nov 29 04:43:25 crc kubenswrapper[4799]: I1129 04:43:25.836997 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7b6db79c5b-mdbvh" Nov 29 04:43:25 crc kubenswrapper[4799]: I1129 04:43:25.855960 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7b6db79c5b-mdbvh" podStartSLOduration=6.855931443 podStartE2EDuration="6.855931443s" podCreationTimestamp="2025-11-29 04:43:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:43:25.855768839 +0000 UTC m=+281.498699239" watchObservedRunningTime="2025-11-29 04:43:25.855931443 +0000 UTC m=+281.498861843" Nov 29 04:43:26 crc kubenswrapper[4799]: I1129 04:43:26.667389 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0fa71a14-00a9-4874-b41c-03254faa648c" path="/var/lib/kubelet/pods/0fa71a14-00a9-4874-b41c-03254faa648c/volumes" Nov 29 04:43:44 crc kubenswrapper[4799]: I1129 04:43:44.518083 4799 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Nov 29 04:43:51 crc kubenswrapper[4799]: I1129 04:43:51.114967 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-xm94x"] Nov 29 04:43:51 crc kubenswrapper[4799]: I1129 04:43:51.117685 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-xm94x" Nov 29 04:43:51 crc kubenswrapper[4799]: I1129 04:43:51.132059 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-xm94x"] Nov 29 04:43:51 crc kubenswrapper[4799]: I1129 04:43:51.165404 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-xm94x\" (UID: \"d6a6b209-8984-4767-b7f7-7d521cefa766\") " pod="openshift-image-registry/image-registry-66df7c8f76-xm94x" Nov 29 04:43:51 crc kubenswrapper[4799]: I1129 04:43:51.165517 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d6a6b209-8984-4767-b7f7-7d521cefa766-installation-pull-secrets\") pod \"image-registry-66df7c8f76-xm94x\" (UID: \"d6a6b209-8984-4767-b7f7-7d521cefa766\") " pod="openshift-image-registry/image-registry-66df7c8f76-xm94x" Nov 29 04:43:51 crc kubenswrapper[4799]: I1129 04:43:51.165571 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d6a6b209-8984-4767-b7f7-7d521cefa766-registry-certificates\") pod \"image-registry-66df7c8f76-xm94x\" (UID: \"d6a6b209-8984-4767-b7f7-7d521cefa766\") " pod="openshift-image-registry/image-registry-66df7c8f76-xm94x" Nov 29 04:43:51 crc kubenswrapper[4799]: I1129 04:43:51.165847 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d6a6b209-8984-4767-b7f7-7d521cefa766-registry-tls\") pod \"image-registry-66df7c8f76-xm94x\" (UID: \"d6a6b209-8984-4767-b7f7-7d521cefa766\") " pod="openshift-image-registry/image-registry-66df7c8f76-xm94x" Nov 29 04:43:51 crc kubenswrapper[4799]: I1129 04:43:51.165950 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d6a6b209-8984-4767-b7f7-7d521cefa766-trusted-ca\") pod \"image-registry-66df7c8f76-xm94x\" (UID: \"d6a6b209-8984-4767-b7f7-7d521cefa766\") " pod="openshift-image-registry/image-registry-66df7c8f76-xm94x" Nov 29 04:43:51 crc kubenswrapper[4799]: I1129 04:43:51.165979 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d6a6b209-8984-4767-b7f7-7d521cefa766-ca-trust-extracted\") pod \"image-registry-66df7c8f76-xm94x\" (UID: \"d6a6b209-8984-4767-b7f7-7d521cefa766\") " pod="openshift-image-registry/image-registry-66df7c8f76-xm94x" Nov 29 04:43:51 crc kubenswrapper[4799]: I1129 04:43:51.167182 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d6a6b209-8984-4767-b7f7-7d521cefa766-bound-sa-token\") pod \"image-registry-66df7c8f76-xm94x\" (UID: \"d6a6b209-8984-4767-b7f7-7d521cefa766\") " pod="openshift-image-registry/image-registry-66df7c8f76-xm94x" Nov 29 04:43:51 crc kubenswrapper[4799]: I1129 04:43:51.167221 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2cz8\" (UniqueName: \"kubernetes.io/projected/d6a6b209-8984-4767-b7f7-7d521cefa766-kube-api-access-r2cz8\") pod \"image-registry-66df7c8f76-xm94x\" (UID: \"d6a6b209-8984-4767-b7f7-7d521cefa766\") " pod="openshift-image-registry/image-registry-66df7c8f76-xm94x" Nov 29 04:43:51 crc kubenswrapper[4799]: I1129 04:43:51.204504 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-xm94x\" (UID: \"d6a6b209-8984-4767-b7f7-7d521cefa766\") " pod="openshift-image-registry/image-registry-66df7c8f76-xm94x" Nov 29 04:43:51 crc kubenswrapper[4799]: I1129 04:43:51.269068 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d6a6b209-8984-4767-b7f7-7d521cefa766-registry-tls\") pod \"image-registry-66df7c8f76-xm94x\" (UID: \"d6a6b209-8984-4767-b7f7-7d521cefa766\") " pod="openshift-image-registry/image-registry-66df7c8f76-xm94x" Nov 29 04:43:51 crc kubenswrapper[4799]: I1129 04:43:51.269133 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d6a6b209-8984-4767-b7f7-7d521cefa766-ca-trust-extracted\") pod \"image-registry-66df7c8f76-xm94x\" (UID: \"d6a6b209-8984-4767-b7f7-7d521cefa766\") " pod="openshift-image-registry/image-registry-66df7c8f76-xm94x" Nov 29 04:43:51 crc kubenswrapper[4799]: I1129 04:43:51.269150 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d6a6b209-8984-4767-b7f7-7d521cefa766-trusted-ca\") pod \"image-registry-66df7c8f76-xm94x\" (UID: \"d6a6b209-8984-4767-b7f7-7d521cefa766\") " pod="openshift-image-registry/image-registry-66df7c8f76-xm94x" Nov 29 04:43:51 crc kubenswrapper[4799]: I1129 04:43:51.269180 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d6a6b209-8984-4767-b7f7-7d521cefa766-bound-sa-token\") pod \"image-registry-66df7c8f76-xm94x\" (UID: \"d6a6b209-8984-4767-b7f7-7d521cefa766\") " pod="openshift-image-registry/image-registry-66df7c8f76-xm94x" Nov 29 04:43:51 crc kubenswrapper[4799]: I1129 04:43:51.269204 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2cz8\" (UniqueName: \"kubernetes.io/projected/d6a6b209-8984-4767-b7f7-7d521cefa766-kube-api-access-r2cz8\") pod \"image-registry-66df7c8f76-xm94x\" (UID: \"d6a6b209-8984-4767-b7f7-7d521cefa766\") " pod="openshift-image-registry/image-registry-66df7c8f76-xm94x" Nov 29 04:43:51 crc kubenswrapper[4799]: I1129 04:43:51.269243 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d6a6b209-8984-4767-b7f7-7d521cefa766-installation-pull-secrets\") pod \"image-registry-66df7c8f76-xm94x\" (UID: \"d6a6b209-8984-4767-b7f7-7d521cefa766\") " pod="openshift-image-registry/image-registry-66df7c8f76-xm94x" Nov 29 04:43:51 crc kubenswrapper[4799]: I1129 04:43:51.269261 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d6a6b209-8984-4767-b7f7-7d521cefa766-registry-certificates\") pod \"image-registry-66df7c8f76-xm94x\" (UID: \"d6a6b209-8984-4767-b7f7-7d521cefa766\") " pod="openshift-image-registry/image-registry-66df7c8f76-xm94x" Nov 29 04:43:51 crc kubenswrapper[4799]: I1129 04:43:51.269959 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d6a6b209-8984-4767-b7f7-7d521cefa766-ca-trust-extracted\") pod \"image-registry-66df7c8f76-xm94x\" (UID: \"d6a6b209-8984-4767-b7f7-7d521cefa766\") " pod="openshift-image-registry/image-registry-66df7c8f76-xm94x" Nov 29 04:43:51 crc kubenswrapper[4799]: I1129 04:43:51.270742 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d6a6b209-8984-4767-b7f7-7d521cefa766-trusted-ca\") pod \"image-registry-66df7c8f76-xm94x\" (UID: \"d6a6b209-8984-4767-b7f7-7d521cefa766\") " pod="openshift-image-registry/image-registry-66df7c8f76-xm94x" Nov 29 04:43:51 crc kubenswrapper[4799]: I1129 04:43:51.270764 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d6a6b209-8984-4767-b7f7-7d521cefa766-registry-certificates\") pod \"image-registry-66df7c8f76-xm94x\" (UID: \"d6a6b209-8984-4767-b7f7-7d521cefa766\") " pod="openshift-image-registry/image-registry-66df7c8f76-xm94x" Nov 29 04:43:51 crc kubenswrapper[4799]: I1129 04:43:51.279532 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d6a6b209-8984-4767-b7f7-7d521cefa766-installation-pull-secrets\") pod \"image-registry-66df7c8f76-xm94x\" (UID: \"d6a6b209-8984-4767-b7f7-7d521cefa766\") " pod="openshift-image-registry/image-registry-66df7c8f76-xm94x" Nov 29 04:43:51 crc kubenswrapper[4799]: I1129 04:43:51.279544 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d6a6b209-8984-4767-b7f7-7d521cefa766-registry-tls\") pod \"image-registry-66df7c8f76-xm94x\" (UID: \"d6a6b209-8984-4767-b7f7-7d521cefa766\") " pod="openshift-image-registry/image-registry-66df7c8f76-xm94x" Nov 29 04:43:51 crc kubenswrapper[4799]: I1129 04:43:51.286515 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d6a6b209-8984-4767-b7f7-7d521cefa766-bound-sa-token\") pod \"image-registry-66df7c8f76-xm94x\" (UID: \"d6a6b209-8984-4767-b7f7-7d521cefa766\") " pod="openshift-image-registry/image-registry-66df7c8f76-xm94x" Nov 29 04:43:51 crc kubenswrapper[4799]: I1129 04:43:51.288482 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2cz8\" (UniqueName: \"kubernetes.io/projected/d6a6b209-8984-4767-b7f7-7d521cefa766-kube-api-access-r2cz8\") pod \"image-registry-66df7c8f76-xm94x\" (UID: \"d6a6b209-8984-4767-b7f7-7d521cefa766\") " pod="openshift-image-registry/image-registry-66df7c8f76-xm94x" Nov 29 04:43:51 crc kubenswrapper[4799]: I1129 04:43:51.440033 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-xm94x" Nov 29 04:43:51 crc kubenswrapper[4799]: I1129 04:43:51.884446 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-xm94x"] Nov 29 04:43:51 crc kubenswrapper[4799]: I1129 04:43:51.990779 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-xm94x" event={"ID":"d6a6b209-8984-4767-b7f7-7d521cefa766","Type":"ContainerStarted","Data":"6efccc1cf4caa563cf3cc7f86954d3a6a4861b12ad7a090726c8a6712c4083e3"} Nov 29 04:43:52 crc kubenswrapper[4799]: I1129 04:43:52.999190 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-xm94x" event={"ID":"d6a6b209-8984-4767-b7f7-7d521cefa766","Type":"ContainerStarted","Data":"0b7199b9e291f7039ead537b679f47457ee9b189bdc0fa9025d275f2e909d2bf"} Nov 29 04:43:53 crc kubenswrapper[4799]: I1129 04:43:52.999369 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-xm94x" Nov 29 04:43:53 crc kubenswrapper[4799]: I1129 04:43:53.026777 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-xm94x" podStartSLOduration=2.026740833 podStartE2EDuration="2.026740833s" podCreationTimestamp="2025-11-29 04:43:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:43:53.024056554 +0000 UTC m=+308.666987004" watchObservedRunningTime="2025-11-29 04:43:53.026740833 +0000 UTC m=+308.669671273" Nov 29 04:44:11 crc kubenswrapper[4799]: I1129 04:44:11.449040 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-xm94x" Nov 29 04:44:11 crc kubenswrapper[4799]: I1129 04:44:11.530337 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-2bbp8"] Nov 29 04:44:36 crc kubenswrapper[4799]: I1129 04:44:36.583176 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" podUID="48c7bbb5-36e2-4677-b259-d3da6302d415" containerName="registry" containerID="cri-o://945cf67f84af377435db6cb0366c38af62992b22dc85a2294b4e1536f1855ae7" gracePeriod=30 Nov 29 04:44:37 crc kubenswrapper[4799]: I1129 04:44:37.089402 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:44:37 crc kubenswrapper[4799]: I1129 04:44:37.226352 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/48c7bbb5-36e2-4677-b259-d3da6302d415-registry-tls\") pod \"48c7bbb5-36e2-4677-b259-d3da6302d415\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " Nov 29 04:44:37 crc kubenswrapper[4799]: I1129 04:44:37.226439 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/48c7bbb5-36e2-4677-b259-d3da6302d415-ca-trust-extracted\") pod \"48c7bbb5-36e2-4677-b259-d3da6302d415\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " Nov 29 04:44:37 crc kubenswrapper[4799]: I1129 04:44:37.226477 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/48c7bbb5-36e2-4677-b259-d3da6302d415-installation-pull-secrets\") pod \"48c7bbb5-36e2-4677-b259-d3da6302d415\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " Nov 29 04:44:37 crc kubenswrapper[4799]: I1129 04:44:37.226532 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/48c7bbb5-36e2-4677-b259-d3da6302d415-trusted-ca\") pod \"48c7bbb5-36e2-4677-b259-d3da6302d415\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " Nov 29 04:44:37 crc kubenswrapper[4799]: I1129 04:44:37.226615 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xvt7f\" (UniqueName: \"kubernetes.io/projected/48c7bbb5-36e2-4677-b259-d3da6302d415-kube-api-access-xvt7f\") pod \"48c7bbb5-36e2-4677-b259-d3da6302d415\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " Nov 29 04:44:37 crc kubenswrapper[4799]: I1129 04:44:37.226683 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/48c7bbb5-36e2-4677-b259-d3da6302d415-bound-sa-token\") pod \"48c7bbb5-36e2-4677-b259-d3da6302d415\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " Nov 29 04:44:37 crc kubenswrapper[4799]: I1129 04:44:37.226718 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/48c7bbb5-36e2-4677-b259-d3da6302d415-registry-certificates\") pod \"48c7bbb5-36e2-4677-b259-d3da6302d415\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " Nov 29 04:44:37 crc kubenswrapper[4799]: I1129 04:44:37.226900 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"48c7bbb5-36e2-4677-b259-d3da6302d415\" (UID: \"48c7bbb5-36e2-4677-b259-d3da6302d415\") " Nov 29 04:44:37 crc kubenswrapper[4799]: I1129 04:44:37.227881 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48c7bbb5-36e2-4677-b259-d3da6302d415-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "48c7bbb5-36e2-4677-b259-d3da6302d415" (UID: "48c7bbb5-36e2-4677-b259-d3da6302d415"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:44:37 crc kubenswrapper[4799]: I1129 04:44:37.227941 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48c7bbb5-36e2-4677-b259-d3da6302d415-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "48c7bbb5-36e2-4677-b259-d3da6302d415" (UID: "48c7bbb5-36e2-4677-b259-d3da6302d415"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:44:37 crc kubenswrapper[4799]: I1129 04:44:37.229571 4799 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/48c7bbb5-36e2-4677-b259-d3da6302d415-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 29 04:44:37 crc kubenswrapper[4799]: I1129 04:44:37.229602 4799 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/48c7bbb5-36e2-4677-b259-d3da6302d415-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 29 04:44:37 crc kubenswrapper[4799]: I1129 04:44:37.234806 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48c7bbb5-36e2-4677-b259-d3da6302d415-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "48c7bbb5-36e2-4677-b259-d3da6302d415" (UID: "48c7bbb5-36e2-4677-b259-d3da6302d415"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:44:37 crc kubenswrapper[4799]: I1129 04:44:37.236119 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48c7bbb5-36e2-4677-b259-d3da6302d415-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "48c7bbb5-36e2-4677-b259-d3da6302d415" (UID: "48c7bbb5-36e2-4677-b259-d3da6302d415"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:44:37 crc kubenswrapper[4799]: I1129 04:44:37.236667 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48c7bbb5-36e2-4677-b259-d3da6302d415-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "48c7bbb5-36e2-4677-b259-d3da6302d415" (UID: "48c7bbb5-36e2-4677-b259-d3da6302d415"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:44:37 crc kubenswrapper[4799]: I1129 04:44:37.237588 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48c7bbb5-36e2-4677-b259-d3da6302d415-kube-api-access-xvt7f" (OuterVolumeSpecName: "kube-api-access-xvt7f") pod "48c7bbb5-36e2-4677-b259-d3da6302d415" (UID: "48c7bbb5-36e2-4677-b259-d3da6302d415"). InnerVolumeSpecName "kube-api-access-xvt7f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:44:37 crc kubenswrapper[4799]: I1129 04:44:37.238712 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "48c7bbb5-36e2-4677-b259-d3da6302d415" (UID: "48c7bbb5-36e2-4677-b259-d3da6302d415"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 29 04:44:37 crc kubenswrapper[4799]: I1129 04:44:37.243334 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48c7bbb5-36e2-4677-b259-d3da6302d415-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "48c7bbb5-36e2-4677-b259-d3da6302d415" (UID: "48c7bbb5-36e2-4677-b259-d3da6302d415"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:44:37 crc kubenswrapper[4799]: I1129 04:44:37.331228 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xvt7f\" (UniqueName: \"kubernetes.io/projected/48c7bbb5-36e2-4677-b259-d3da6302d415-kube-api-access-xvt7f\") on node \"crc\" DevicePath \"\"" Nov 29 04:44:37 crc kubenswrapper[4799]: I1129 04:44:37.331343 4799 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/48c7bbb5-36e2-4677-b259-d3da6302d415-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 29 04:44:37 crc kubenswrapper[4799]: I1129 04:44:37.331357 4799 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/48c7bbb5-36e2-4677-b259-d3da6302d415-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 29 04:44:37 crc kubenswrapper[4799]: I1129 04:44:37.331371 4799 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/48c7bbb5-36e2-4677-b259-d3da6302d415-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 29 04:44:37 crc kubenswrapper[4799]: I1129 04:44:37.331388 4799 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/48c7bbb5-36e2-4677-b259-d3da6302d415-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 29 04:44:37 crc kubenswrapper[4799]: I1129 04:44:37.355452 4799 generic.go:334] "Generic (PLEG): container finished" podID="48c7bbb5-36e2-4677-b259-d3da6302d415" containerID="945cf67f84af377435db6cb0366c38af62992b22dc85a2294b4e1536f1855ae7" exitCode=0 Nov 29 04:44:37 crc kubenswrapper[4799]: I1129 04:44:37.355545 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" event={"ID":"48c7bbb5-36e2-4677-b259-d3da6302d415","Type":"ContainerDied","Data":"945cf67f84af377435db6cb0366c38af62992b22dc85a2294b4e1536f1855ae7"} Nov 29 04:44:37 crc kubenswrapper[4799]: I1129 04:44:37.355588 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" Nov 29 04:44:37 crc kubenswrapper[4799]: I1129 04:44:37.355617 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-2bbp8" event={"ID":"48c7bbb5-36e2-4677-b259-d3da6302d415","Type":"ContainerDied","Data":"50327e8968746366dd798385527fba10ff6ff8d92c9618a20547417e602a4d54"} Nov 29 04:44:37 crc kubenswrapper[4799]: I1129 04:44:37.355667 4799 scope.go:117] "RemoveContainer" containerID="945cf67f84af377435db6cb0366c38af62992b22dc85a2294b4e1536f1855ae7" Nov 29 04:44:37 crc kubenswrapper[4799]: I1129 04:44:37.380925 4799 scope.go:117] "RemoveContainer" containerID="945cf67f84af377435db6cb0366c38af62992b22dc85a2294b4e1536f1855ae7" Nov 29 04:44:37 crc kubenswrapper[4799]: E1129 04:44:37.381782 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"945cf67f84af377435db6cb0366c38af62992b22dc85a2294b4e1536f1855ae7\": container with ID starting with 945cf67f84af377435db6cb0366c38af62992b22dc85a2294b4e1536f1855ae7 not found: ID does not exist" containerID="945cf67f84af377435db6cb0366c38af62992b22dc85a2294b4e1536f1855ae7" Nov 29 04:44:37 crc kubenswrapper[4799]: I1129 04:44:37.381872 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"945cf67f84af377435db6cb0366c38af62992b22dc85a2294b4e1536f1855ae7"} err="failed to get container status \"945cf67f84af377435db6cb0366c38af62992b22dc85a2294b4e1536f1855ae7\": rpc error: code = NotFound desc = could not find container \"945cf67f84af377435db6cb0366c38af62992b22dc85a2294b4e1536f1855ae7\": container with ID starting with 945cf67f84af377435db6cb0366c38af62992b22dc85a2294b4e1536f1855ae7 not found: ID does not exist" Nov 29 04:44:37 crc kubenswrapper[4799]: I1129 04:44:37.401874 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-2bbp8"] Nov 29 04:44:37 crc kubenswrapper[4799]: I1129 04:44:37.407736 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-2bbp8"] Nov 29 04:44:37 crc kubenswrapper[4799]: I1129 04:44:37.498153 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 04:44:37 crc kubenswrapper[4799]: I1129 04:44:37.498282 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 04:44:38 crc kubenswrapper[4799]: I1129 04:44:38.668471 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48c7bbb5-36e2-4677-b259-d3da6302d415" path="/var/lib/kubelet/pods/48c7bbb5-36e2-4677-b259-d3da6302d415/volumes" Nov 29 04:45:00 crc kubenswrapper[4799]: I1129 04:45:00.188542 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406525-md8rr"] Nov 29 04:45:00 crc kubenswrapper[4799]: E1129 04:45:00.190686 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48c7bbb5-36e2-4677-b259-d3da6302d415" containerName="registry" Nov 29 04:45:00 crc kubenswrapper[4799]: I1129 04:45:00.190828 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="48c7bbb5-36e2-4677-b259-d3da6302d415" containerName="registry" Nov 29 04:45:00 crc kubenswrapper[4799]: I1129 04:45:00.191028 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="48c7bbb5-36e2-4677-b259-d3da6302d415" containerName="registry" Nov 29 04:45:00 crc kubenswrapper[4799]: I1129 04:45:00.191594 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406525-md8rr" Nov 29 04:45:00 crc kubenswrapper[4799]: I1129 04:45:00.195153 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 29 04:45:00 crc kubenswrapper[4799]: I1129 04:45:00.198348 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 29 04:45:00 crc kubenswrapper[4799]: I1129 04:45:00.202577 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406525-md8rr"] Nov 29 04:45:00 crc kubenswrapper[4799]: I1129 04:45:00.269002 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c614b62c-cd99-4f49-8939-25fe713e4809-config-volume\") pod \"collect-profiles-29406525-md8rr\" (UID: \"c614b62c-cd99-4f49-8939-25fe713e4809\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406525-md8rr" Nov 29 04:45:00 crc kubenswrapper[4799]: I1129 04:45:00.269120 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hj54v\" (UniqueName: \"kubernetes.io/projected/c614b62c-cd99-4f49-8939-25fe713e4809-kube-api-access-hj54v\") pod \"collect-profiles-29406525-md8rr\" (UID: \"c614b62c-cd99-4f49-8939-25fe713e4809\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406525-md8rr" Nov 29 04:45:00 crc kubenswrapper[4799]: I1129 04:45:00.269174 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c614b62c-cd99-4f49-8939-25fe713e4809-secret-volume\") pod \"collect-profiles-29406525-md8rr\" (UID: \"c614b62c-cd99-4f49-8939-25fe713e4809\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406525-md8rr" Nov 29 04:45:00 crc kubenswrapper[4799]: I1129 04:45:00.372048 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c614b62c-cd99-4f49-8939-25fe713e4809-config-volume\") pod \"collect-profiles-29406525-md8rr\" (UID: \"c614b62c-cd99-4f49-8939-25fe713e4809\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406525-md8rr" Nov 29 04:45:00 crc kubenswrapper[4799]: I1129 04:45:00.372656 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hj54v\" (UniqueName: \"kubernetes.io/projected/c614b62c-cd99-4f49-8939-25fe713e4809-kube-api-access-hj54v\") pod \"collect-profiles-29406525-md8rr\" (UID: \"c614b62c-cd99-4f49-8939-25fe713e4809\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406525-md8rr" Nov 29 04:45:00 crc kubenswrapper[4799]: I1129 04:45:00.372916 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c614b62c-cd99-4f49-8939-25fe713e4809-secret-volume\") pod \"collect-profiles-29406525-md8rr\" (UID: \"c614b62c-cd99-4f49-8939-25fe713e4809\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406525-md8rr" Nov 29 04:45:00 crc kubenswrapper[4799]: I1129 04:45:00.374096 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c614b62c-cd99-4f49-8939-25fe713e4809-config-volume\") pod \"collect-profiles-29406525-md8rr\" (UID: \"c614b62c-cd99-4f49-8939-25fe713e4809\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406525-md8rr" Nov 29 04:45:00 crc kubenswrapper[4799]: I1129 04:45:00.379832 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c614b62c-cd99-4f49-8939-25fe713e4809-secret-volume\") pod \"collect-profiles-29406525-md8rr\" (UID: \"c614b62c-cd99-4f49-8939-25fe713e4809\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406525-md8rr" Nov 29 04:45:00 crc kubenswrapper[4799]: I1129 04:45:00.397327 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hj54v\" (UniqueName: \"kubernetes.io/projected/c614b62c-cd99-4f49-8939-25fe713e4809-kube-api-access-hj54v\") pod \"collect-profiles-29406525-md8rr\" (UID: \"c614b62c-cd99-4f49-8939-25fe713e4809\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406525-md8rr" Nov 29 04:45:00 crc kubenswrapper[4799]: I1129 04:45:00.519010 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406525-md8rr" Nov 29 04:45:00 crc kubenswrapper[4799]: I1129 04:45:00.934559 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406525-md8rr"] Nov 29 04:45:01 crc kubenswrapper[4799]: I1129 04:45:01.553721 4799 generic.go:334] "Generic (PLEG): container finished" podID="c614b62c-cd99-4f49-8939-25fe713e4809" containerID="2abf20d02fcbf7ac03d4005350466966763796841a92dd495d6066c8ab273a23" exitCode=0 Nov 29 04:45:01 crc kubenswrapper[4799]: I1129 04:45:01.553839 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29406525-md8rr" event={"ID":"c614b62c-cd99-4f49-8939-25fe713e4809","Type":"ContainerDied","Data":"2abf20d02fcbf7ac03d4005350466966763796841a92dd495d6066c8ab273a23"} Nov 29 04:45:01 crc kubenswrapper[4799]: I1129 04:45:01.554182 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29406525-md8rr" event={"ID":"c614b62c-cd99-4f49-8939-25fe713e4809","Type":"ContainerStarted","Data":"2e159de5e3488d5ec7f95166f7d312eb91687fb9f2e4bd041ec60bdbcceb6ef6"} Nov 29 04:45:02 crc kubenswrapper[4799]: I1129 04:45:02.859521 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406525-md8rr" Nov 29 04:45:02 crc kubenswrapper[4799]: I1129 04:45:02.907590 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c614b62c-cd99-4f49-8939-25fe713e4809-config-volume\") pod \"c614b62c-cd99-4f49-8939-25fe713e4809\" (UID: \"c614b62c-cd99-4f49-8939-25fe713e4809\") " Nov 29 04:45:02 crc kubenswrapper[4799]: I1129 04:45:02.907729 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c614b62c-cd99-4f49-8939-25fe713e4809-secret-volume\") pod \"c614b62c-cd99-4f49-8939-25fe713e4809\" (UID: \"c614b62c-cd99-4f49-8939-25fe713e4809\") " Nov 29 04:45:02 crc kubenswrapper[4799]: I1129 04:45:02.907754 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hj54v\" (UniqueName: \"kubernetes.io/projected/c614b62c-cd99-4f49-8939-25fe713e4809-kube-api-access-hj54v\") pod \"c614b62c-cd99-4f49-8939-25fe713e4809\" (UID: \"c614b62c-cd99-4f49-8939-25fe713e4809\") " Nov 29 04:45:02 crc kubenswrapper[4799]: I1129 04:45:02.908394 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c614b62c-cd99-4f49-8939-25fe713e4809-config-volume" (OuterVolumeSpecName: "config-volume") pod "c614b62c-cd99-4f49-8939-25fe713e4809" (UID: "c614b62c-cd99-4f49-8939-25fe713e4809"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:45:02 crc kubenswrapper[4799]: I1129 04:45:02.916126 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c614b62c-cd99-4f49-8939-25fe713e4809-kube-api-access-hj54v" (OuterVolumeSpecName: "kube-api-access-hj54v") pod "c614b62c-cd99-4f49-8939-25fe713e4809" (UID: "c614b62c-cd99-4f49-8939-25fe713e4809"). InnerVolumeSpecName "kube-api-access-hj54v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:45:02 crc kubenswrapper[4799]: I1129 04:45:02.916317 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c614b62c-cd99-4f49-8939-25fe713e4809-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c614b62c-cd99-4f49-8939-25fe713e4809" (UID: "c614b62c-cd99-4f49-8939-25fe713e4809"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:45:03 crc kubenswrapper[4799]: I1129 04:45:03.009905 4799 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c614b62c-cd99-4f49-8939-25fe713e4809-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 29 04:45:03 crc kubenswrapper[4799]: I1129 04:45:03.009969 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hj54v\" (UniqueName: \"kubernetes.io/projected/c614b62c-cd99-4f49-8939-25fe713e4809-kube-api-access-hj54v\") on node \"crc\" DevicePath \"\"" Nov 29 04:45:03 crc kubenswrapper[4799]: I1129 04:45:03.009984 4799 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c614b62c-cd99-4f49-8939-25fe713e4809-config-volume\") on node \"crc\" DevicePath \"\"" Nov 29 04:45:03 crc kubenswrapper[4799]: I1129 04:45:03.571881 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29406525-md8rr" event={"ID":"c614b62c-cd99-4f49-8939-25fe713e4809","Type":"ContainerDied","Data":"2e159de5e3488d5ec7f95166f7d312eb91687fb9f2e4bd041ec60bdbcceb6ef6"} Nov 29 04:45:03 crc kubenswrapper[4799]: I1129 04:45:03.571959 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2e159de5e3488d5ec7f95166f7d312eb91687fb9f2e4bd041ec60bdbcceb6ef6" Nov 29 04:45:03 crc kubenswrapper[4799]: I1129 04:45:03.572461 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406525-md8rr" Nov 29 04:45:07 crc kubenswrapper[4799]: I1129 04:45:07.498520 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 04:45:07 crc kubenswrapper[4799]: I1129 04:45:07.499098 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 04:45:37 crc kubenswrapper[4799]: I1129 04:45:37.498650 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 04:45:37 crc kubenswrapper[4799]: I1129 04:45:37.499548 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 04:45:37 crc kubenswrapper[4799]: I1129 04:45:37.499628 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-26t88" Nov 29 04:45:37 crc kubenswrapper[4799]: I1129 04:45:37.500693 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0ae025ef68b62241dcb107a107b30c8479869f4f96294d65af8cc7c29fb998ff"} pod="openshift-machine-config-operator/machine-config-daemon-26t88" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 29 04:45:37 crc kubenswrapper[4799]: I1129 04:45:37.500765 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" containerID="cri-o://0ae025ef68b62241dcb107a107b30c8479869f4f96294d65af8cc7c29fb998ff" gracePeriod=600 Nov 29 04:45:37 crc kubenswrapper[4799]: I1129 04:45:37.799349 4799 generic.go:334] "Generic (PLEG): container finished" podID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerID="0ae025ef68b62241dcb107a107b30c8479869f4f96294d65af8cc7c29fb998ff" exitCode=0 Nov 29 04:45:37 crc kubenswrapper[4799]: I1129 04:45:37.799439 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" event={"ID":"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8","Type":"ContainerDied","Data":"0ae025ef68b62241dcb107a107b30c8479869f4f96294d65af8cc7c29fb998ff"} Nov 29 04:45:37 crc kubenswrapper[4799]: I1129 04:45:37.800603 4799 scope.go:117] "RemoveContainer" containerID="83b2ec8f0d384066ba90eb42c1a0a925ade3722b2515b0592d6f2a8f5e533a0f" Nov 29 04:45:38 crc kubenswrapper[4799]: I1129 04:45:38.812046 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" event={"ID":"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8","Type":"ContainerStarted","Data":"2b1b7386c0bb5e909eb4074926c911c29f96ed91593e4e76b3689a1397aeb0ff"} Nov 29 04:47:37 crc kubenswrapper[4799]: I1129 04:47:37.497746 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 04:47:37 crc kubenswrapper[4799]: I1129 04:47:37.498626 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 04:47:45 crc kubenswrapper[4799]: I1129 04:47:45.465843 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-68jsj"] Nov 29 04:47:45 crc kubenswrapper[4799]: E1129 04:47:45.466612 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c614b62c-cd99-4f49-8939-25fe713e4809" containerName="collect-profiles" Nov 29 04:47:45 crc kubenswrapper[4799]: I1129 04:47:45.466628 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="c614b62c-cd99-4f49-8939-25fe713e4809" containerName="collect-profiles" Nov 29 04:47:45 crc kubenswrapper[4799]: I1129 04:47:45.466725 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="c614b62c-cd99-4f49-8939-25fe713e4809" containerName="collect-profiles" Nov 29 04:47:45 crc kubenswrapper[4799]: I1129 04:47:45.467210 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-68jsj" Nov 29 04:47:45 crc kubenswrapper[4799]: I1129 04:47:45.472019 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Nov 29 04:47:45 crc kubenswrapper[4799]: I1129 04:47:45.472084 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Nov 29 04:47:45 crc kubenswrapper[4799]: I1129 04:47:45.472085 4799 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-cbzsf" Nov 29 04:47:45 crc kubenswrapper[4799]: I1129 04:47:45.478418 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-4hw5r"] Nov 29 04:47:45 crc kubenswrapper[4799]: I1129 04:47:45.479404 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-4hw5r" Nov 29 04:47:45 crc kubenswrapper[4799]: I1129 04:47:45.484508 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-68jsj"] Nov 29 04:47:45 crc kubenswrapper[4799]: I1129 04:47:45.484528 4799 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-q2555" Nov 29 04:47:45 crc kubenswrapper[4799]: I1129 04:47:45.497855 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-4hw5r"] Nov 29 04:47:45 crc kubenswrapper[4799]: I1129 04:47:45.502272 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-b9lmh"] Nov 29 04:47:45 crc kubenswrapper[4799]: I1129 04:47:45.503221 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-b9lmh" Nov 29 04:47:45 crc kubenswrapper[4799]: I1129 04:47:45.507901 4799 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-q7hg6" Nov 29 04:47:45 crc kubenswrapper[4799]: I1129 04:47:45.520432 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-b9lmh"] Nov 29 04:47:45 crc kubenswrapper[4799]: I1129 04:47:45.647937 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjq7m\" (UniqueName: \"kubernetes.io/projected/90513767-7b60-461e-bb69-727a8f25e15e-kube-api-access-wjq7m\") pod \"cert-manager-cainjector-7f985d654d-68jsj\" (UID: \"90513767-7b60-461e-bb69-727a8f25e15e\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-68jsj" Nov 29 04:47:45 crc kubenswrapper[4799]: I1129 04:47:45.648233 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7g9jl\" (UniqueName: \"kubernetes.io/projected/9347b23b-a5d2-40f4-8d55-a320f5a0c78a-kube-api-access-7g9jl\") pod \"cert-manager-5b446d88c5-4hw5r\" (UID: \"9347b23b-a5d2-40f4-8d55-a320f5a0c78a\") " pod="cert-manager/cert-manager-5b446d88c5-4hw5r" Nov 29 04:47:45 crc kubenswrapper[4799]: I1129 04:47:45.648265 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jcvg\" (UniqueName: \"kubernetes.io/projected/28341ac5-a271-4435-a855-512a548beb96-kube-api-access-9jcvg\") pod \"cert-manager-webhook-5655c58dd6-b9lmh\" (UID: \"28341ac5-a271-4435-a855-512a548beb96\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-b9lmh" Nov 29 04:47:45 crc kubenswrapper[4799]: I1129 04:47:45.750662 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7g9jl\" (UniqueName: \"kubernetes.io/projected/9347b23b-a5d2-40f4-8d55-a320f5a0c78a-kube-api-access-7g9jl\") pod \"cert-manager-5b446d88c5-4hw5r\" (UID: \"9347b23b-a5d2-40f4-8d55-a320f5a0c78a\") " pod="cert-manager/cert-manager-5b446d88c5-4hw5r" Nov 29 04:47:45 crc kubenswrapper[4799]: I1129 04:47:45.750719 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jcvg\" (UniqueName: \"kubernetes.io/projected/28341ac5-a271-4435-a855-512a548beb96-kube-api-access-9jcvg\") pod \"cert-manager-webhook-5655c58dd6-b9lmh\" (UID: \"28341ac5-a271-4435-a855-512a548beb96\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-b9lmh" Nov 29 04:47:45 crc kubenswrapper[4799]: I1129 04:47:45.750826 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjq7m\" (UniqueName: \"kubernetes.io/projected/90513767-7b60-461e-bb69-727a8f25e15e-kube-api-access-wjq7m\") pod \"cert-manager-cainjector-7f985d654d-68jsj\" (UID: \"90513767-7b60-461e-bb69-727a8f25e15e\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-68jsj" Nov 29 04:47:45 crc kubenswrapper[4799]: I1129 04:47:45.771504 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjq7m\" (UniqueName: \"kubernetes.io/projected/90513767-7b60-461e-bb69-727a8f25e15e-kube-api-access-wjq7m\") pod \"cert-manager-cainjector-7f985d654d-68jsj\" (UID: \"90513767-7b60-461e-bb69-727a8f25e15e\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-68jsj" Nov 29 04:47:45 crc kubenswrapper[4799]: I1129 04:47:45.771504 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jcvg\" (UniqueName: \"kubernetes.io/projected/28341ac5-a271-4435-a855-512a548beb96-kube-api-access-9jcvg\") pod \"cert-manager-webhook-5655c58dd6-b9lmh\" (UID: \"28341ac5-a271-4435-a855-512a548beb96\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-b9lmh" Nov 29 04:47:45 crc kubenswrapper[4799]: I1129 04:47:45.772220 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7g9jl\" (UniqueName: \"kubernetes.io/projected/9347b23b-a5d2-40f4-8d55-a320f5a0c78a-kube-api-access-7g9jl\") pod \"cert-manager-5b446d88c5-4hw5r\" (UID: \"9347b23b-a5d2-40f4-8d55-a320f5a0c78a\") " pod="cert-manager/cert-manager-5b446d88c5-4hw5r" Nov 29 04:47:45 crc kubenswrapper[4799]: I1129 04:47:45.785637 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-68jsj" Nov 29 04:47:45 crc kubenswrapper[4799]: I1129 04:47:45.798529 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-4hw5r" Nov 29 04:47:45 crc kubenswrapper[4799]: I1129 04:47:45.825156 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-b9lmh" Nov 29 04:47:46 crc kubenswrapper[4799]: I1129 04:47:46.245995 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-68jsj"] Nov 29 04:47:46 crc kubenswrapper[4799]: W1129 04:47:46.252331 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod90513767_7b60_461e_bb69_727a8f25e15e.slice/crio-188ec9e716c3d6cd996d48c89700a65b5c8e409cb3e7dd98ecaf8d400d8030c4 WatchSource:0}: Error finding container 188ec9e716c3d6cd996d48c89700a65b5c8e409cb3e7dd98ecaf8d400d8030c4: Status 404 returned error can't find the container with id 188ec9e716c3d6cd996d48c89700a65b5c8e409cb3e7dd98ecaf8d400d8030c4 Nov 29 04:47:46 crc kubenswrapper[4799]: I1129 04:47:46.255905 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 29 04:47:46 crc kubenswrapper[4799]: I1129 04:47:46.295939 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-4hw5r"] Nov 29 04:47:46 crc kubenswrapper[4799]: I1129 04:47:46.299337 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-b9lmh"] Nov 29 04:47:46 crc kubenswrapper[4799]: W1129 04:47:46.305215 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9347b23b_a5d2_40f4_8d55_a320f5a0c78a.slice/crio-a58c8523411d4373a193115cf3a492ff4f9d4ec622281c5cbdf1c09cafbcdd8b WatchSource:0}: Error finding container a58c8523411d4373a193115cf3a492ff4f9d4ec622281c5cbdf1c09cafbcdd8b: Status 404 returned error can't find the container with id a58c8523411d4373a193115cf3a492ff4f9d4ec622281c5cbdf1c09cafbcdd8b Nov 29 04:47:46 crc kubenswrapper[4799]: I1129 04:47:46.723856 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-4hw5r" event={"ID":"9347b23b-a5d2-40f4-8d55-a320f5a0c78a","Type":"ContainerStarted","Data":"a58c8523411d4373a193115cf3a492ff4f9d4ec622281c5cbdf1c09cafbcdd8b"} Nov 29 04:47:46 crc kubenswrapper[4799]: I1129 04:47:46.725546 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-b9lmh" event={"ID":"28341ac5-a271-4435-a855-512a548beb96","Type":"ContainerStarted","Data":"a605f6b4e755196bda2a062e3d34a4dab939ab83f3b6a8de57052f1538f413b5"} Nov 29 04:47:46 crc kubenswrapper[4799]: I1129 04:47:46.727848 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-68jsj" event={"ID":"90513767-7b60-461e-bb69-727a8f25e15e","Type":"ContainerStarted","Data":"188ec9e716c3d6cd996d48c89700a65b5c8e409cb3e7dd98ecaf8d400d8030c4"} Nov 29 04:47:51 crc kubenswrapper[4799]: I1129 04:47:51.760626 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-b9lmh" event={"ID":"28341ac5-a271-4435-a855-512a548beb96","Type":"ContainerStarted","Data":"39ca2270f715743dca08f208d9ad3da7b89062aa839066e9ab8e0f1d7aa9d34e"} Nov 29 04:47:51 crc kubenswrapper[4799]: I1129 04:47:51.762534 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-b9lmh" Nov 29 04:47:51 crc kubenswrapper[4799]: I1129 04:47:51.764675 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-68jsj" event={"ID":"90513767-7b60-461e-bb69-727a8f25e15e","Type":"ContainerStarted","Data":"c396530a87cebf742b90b3a9c8d37a0648bda98f578f9b8cfb6ce632cd4bae8a"} Nov 29 04:47:51 crc kubenswrapper[4799]: I1129 04:47:51.766962 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-4hw5r" event={"ID":"9347b23b-a5d2-40f4-8d55-a320f5a0c78a","Type":"ContainerStarted","Data":"1728d6173fb549f765c3ce696f2e39ddc0194ea20771c227f86665caf597c702"} Nov 29 04:47:51 crc kubenswrapper[4799]: I1129 04:47:51.779506 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-b9lmh" podStartSLOduration=2.450068269 podStartE2EDuration="6.779472788s" podCreationTimestamp="2025-11-29 04:47:45 +0000 UTC" firstStartedPulling="2025-11-29 04:47:46.305215341 +0000 UTC m=+541.948145741" lastFinishedPulling="2025-11-29 04:47:50.63461986 +0000 UTC m=+546.277550260" observedRunningTime="2025-11-29 04:47:51.776566404 +0000 UTC m=+547.419496804" watchObservedRunningTime="2025-11-29 04:47:51.779472788 +0000 UTC m=+547.422403228" Nov 29 04:47:51 crc kubenswrapper[4799]: I1129 04:47:51.798913 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-68jsj" podStartSLOduration=2.485625936 podStartE2EDuration="6.798876101s" podCreationTimestamp="2025-11-29 04:47:45 +0000 UTC" firstStartedPulling="2025-11-29 04:47:46.255582315 +0000 UTC m=+541.898512715" lastFinishedPulling="2025-11-29 04:47:50.56883248 +0000 UTC m=+546.211762880" observedRunningTime="2025-11-29 04:47:51.793101085 +0000 UTC m=+547.436031525" watchObservedRunningTime="2025-11-29 04:47:51.798876101 +0000 UTC m=+547.441806511" Nov 29 04:47:51 crc kubenswrapper[4799]: I1129 04:47:51.818396 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-4hw5r" podStartSLOduration=2.6334830719999998 podStartE2EDuration="6.818356096s" podCreationTimestamp="2025-11-29 04:47:45 +0000 UTC" firstStartedPulling="2025-11-29 04:47:46.309446078 +0000 UTC m=+541.952376478" lastFinishedPulling="2025-11-29 04:47:50.494319072 +0000 UTC m=+546.137249502" observedRunningTime="2025-11-29 04:47:51.811141665 +0000 UTC m=+547.454072075" watchObservedRunningTime="2025-11-29 04:47:51.818356096 +0000 UTC m=+547.461286496" Nov 29 04:47:55 crc kubenswrapper[4799]: I1129 04:47:55.830297 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-b9lmh" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.147367 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-fht8l"] Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.147907 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" podUID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerName="ovn-controller" containerID="cri-o://8e76d91c4f461a1912117cb8ed3a8087c5efaaf0a67455b721bfbeaad1c7dda8" gracePeriod=30 Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.147972 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" podUID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerName="sbdb" containerID="cri-o://d135e42b8886f1f118abf8f4ce4cf2875ac4dc6c148bfc0677cef6b3651562c0" gracePeriod=30 Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.148039 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" podUID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://dac1c084955d874741d36709fa848c2c475a2d7108d4378b677bb13a88184105" gracePeriod=30 Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.148172 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" podUID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerName="northd" containerID="cri-o://0c81d94f098efec48c1ff8e172e476bf806ccec5777974eea9c0f3850e5fd40e" gracePeriod=30 Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.148064 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" podUID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerName="kube-rbac-proxy-node" containerID="cri-o://28499112c33332c23be39a11cdf7fc4e1946d04524a3f7c534087179a6620121" gracePeriod=30 Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.148197 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" podUID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerName="nbdb" containerID="cri-o://72802532a47b381ead2507e1ae31c49852f9b1d332f3c08fb1e3a972a128eead" gracePeriod=30 Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.148081 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" podUID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerName="ovn-acl-logging" containerID="cri-o://e3d7f1ba1a408c54f0470fca40e307b70ba208dc7eeece599b684820e3f56c41" gracePeriod=30 Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.190258 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" podUID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerName="ovnkube-controller" containerID="cri-o://66dd6f66933db1fc65c8aa5ac9ce4117213e4a0a3f1260975e0d9a6e06366f3d" gracePeriod=30 Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.808166 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fht8l_bd007c61-4689-4179-95ad-44a9724e8e7e/ovnkube-controller/2.log" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.812576 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fht8l_bd007c61-4689-4179-95ad-44a9724e8e7e/ovn-acl-logging/0.log" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.813453 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fht8l_bd007c61-4689-4179-95ad-44a9724e8e7e/ovn-controller/0.log" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.814993 4799 generic.go:334] "Generic (PLEG): container finished" podID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerID="66dd6f66933db1fc65c8aa5ac9ce4117213e4a0a3f1260975e0d9a6e06366f3d" exitCode=0 Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.815042 4799 generic.go:334] "Generic (PLEG): container finished" podID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerID="d135e42b8886f1f118abf8f4ce4cf2875ac4dc6c148bfc0677cef6b3651562c0" exitCode=0 Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.815068 4799 generic.go:334] "Generic (PLEG): container finished" podID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerID="72802532a47b381ead2507e1ae31c49852f9b1d332f3c08fb1e3a972a128eead" exitCode=0 Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.815087 4799 generic.go:334] "Generic (PLEG): container finished" podID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerID="0c81d94f098efec48c1ff8e172e476bf806ccec5777974eea9c0f3850e5fd40e" exitCode=0 Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.815104 4799 generic.go:334] "Generic (PLEG): container finished" podID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerID="dac1c084955d874741d36709fa848c2c475a2d7108d4378b677bb13a88184105" exitCode=0 Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.815089 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" event={"ID":"bd007c61-4689-4179-95ad-44a9724e8e7e","Type":"ContainerDied","Data":"66dd6f66933db1fc65c8aa5ac9ce4117213e4a0a3f1260975e0d9a6e06366f3d"} Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.815180 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" event={"ID":"bd007c61-4689-4179-95ad-44a9724e8e7e","Type":"ContainerDied","Data":"d135e42b8886f1f118abf8f4ce4cf2875ac4dc6c148bfc0677cef6b3651562c0"} Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.815206 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" event={"ID":"bd007c61-4689-4179-95ad-44a9724e8e7e","Type":"ContainerDied","Data":"72802532a47b381ead2507e1ae31c49852f9b1d332f3c08fb1e3a972a128eead"} Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.815228 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" event={"ID":"bd007c61-4689-4179-95ad-44a9724e8e7e","Type":"ContainerDied","Data":"0c81d94f098efec48c1ff8e172e476bf806ccec5777974eea9c0f3850e5fd40e"} Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.815255 4799 scope.go:117] "RemoveContainer" containerID="d5ce3c68997bc5859df929b844d823adf00352454bb278d57e90edcc44817cbe" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.815264 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" event={"ID":"bd007c61-4689-4179-95ad-44a9724e8e7e","Type":"ContainerDied","Data":"dac1c084955d874741d36709fa848c2c475a2d7108d4378b677bb13a88184105"} Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.815296 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" event={"ID":"bd007c61-4689-4179-95ad-44a9724e8e7e","Type":"ContainerDied","Data":"28499112c33332c23be39a11cdf7fc4e1946d04524a3f7c534087179a6620121"} Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.815121 4799 generic.go:334] "Generic (PLEG): container finished" podID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerID="28499112c33332c23be39a11cdf7fc4e1946d04524a3f7c534087179a6620121" exitCode=0 Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.815354 4799 generic.go:334] "Generic (PLEG): container finished" podID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerID="e3d7f1ba1a408c54f0470fca40e307b70ba208dc7eeece599b684820e3f56c41" exitCode=143 Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.815393 4799 generic.go:334] "Generic (PLEG): container finished" podID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerID="8e76d91c4f461a1912117cb8ed3a8087c5efaaf0a67455b721bfbeaad1c7dda8" exitCode=143 Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.815397 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" event={"ID":"bd007c61-4689-4179-95ad-44a9724e8e7e","Type":"ContainerDied","Data":"e3d7f1ba1a408c54f0470fca40e307b70ba208dc7eeece599b684820e3f56c41"} Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.815462 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" event={"ID":"bd007c61-4689-4179-95ad-44a9724e8e7e","Type":"ContainerDied","Data":"8e76d91c4f461a1912117cb8ed3a8087c5efaaf0a67455b721bfbeaad1c7dda8"} Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.818497 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-qt2dl_3b6c9236-5f7d-46c5-b080-4667620d874e/kube-multus/1.log" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.819177 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-qt2dl_3b6c9236-5f7d-46c5-b080-4667620d874e/kube-multus/0.log" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.819229 4799 generic.go:334] "Generic (PLEG): container finished" podID="3b6c9236-5f7d-46c5-b080-4667620d874e" containerID="d0e3c7d98a4a048822bc263c302eef3b130a53051289505324e512944f332b5d" exitCode=2 Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.819261 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-qt2dl" event={"ID":"3b6c9236-5f7d-46c5-b080-4667620d874e","Type":"ContainerDied","Data":"d0e3c7d98a4a048822bc263c302eef3b130a53051289505324e512944f332b5d"} Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.819863 4799 scope.go:117] "RemoveContainer" containerID="d0e3c7d98a4a048822bc263c302eef3b130a53051289505324e512944f332b5d" Nov 29 04:47:56 crc kubenswrapper[4799]: E1129 04:47:56.820280 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-qt2dl_openshift-multus(3b6c9236-5f7d-46c5-b080-4667620d874e)\"" pod="openshift-multus/multus-qt2dl" podUID="3b6c9236-5f7d-46c5-b080-4667620d874e" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.858336 4799 scope.go:117] "RemoveContainer" containerID="92b6c827d0c57a87808fef301661fce28c50359e85c9e47947c50945f4f869de" Nov 29 04:47:56 crc kubenswrapper[4799]: E1129 04:47:56.860070 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5ce3c68997bc5859df929b844d823adf00352454bb278d57e90edcc44817cbe\": container with ID starting with d5ce3c68997bc5859df929b844d823adf00352454bb278d57e90edcc44817cbe not found: ID does not exist" containerID="d5ce3c68997bc5859df929b844d823adf00352454bb278d57e90edcc44817cbe" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.866282 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fht8l_bd007c61-4689-4179-95ad-44a9724e8e7e/ovn-acl-logging/0.log" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.868514 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fht8l_bd007c61-4689-4179-95ad-44a9724e8e7e/ovn-controller/0.log" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.869538 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.924278 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-etc-openvswitch\") pod \"bd007c61-4689-4179-95ad-44a9724e8e7e\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.924359 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-host-cni-bin\") pod \"bd007c61-4689-4179-95ad-44a9724e8e7e\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.924393 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-log-socket\") pod \"bd007c61-4689-4179-95ad-44a9724e8e7e\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.924454 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-run-openvswitch\") pod \"bd007c61-4689-4179-95ad-44a9724e8e7e\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.924477 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-systemd-units\") pod \"bd007c61-4689-4179-95ad-44a9724e8e7e\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.924499 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-host-run-ovn-kubernetes\") pod \"bd007c61-4689-4179-95ad-44a9724e8e7e\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.924532 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bd007c61-4689-4179-95ad-44a9724e8e7e-env-overrides\") pod \"bd007c61-4689-4179-95ad-44a9724e8e7e\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.924572 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-run-systemd\") pod \"bd007c61-4689-4179-95ad-44a9724e8e7e\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.924601 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jpd4j\" (UniqueName: \"kubernetes.io/projected/bd007c61-4689-4179-95ad-44a9724e8e7e-kube-api-access-jpd4j\") pod \"bd007c61-4689-4179-95ad-44a9724e8e7e\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.924629 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bd007c61-4689-4179-95ad-44a9724e8e7e-ovn-node-metrics-cert\") pod \"bd007c61-4689-4179-95ad-44a9724e8e7e\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.924648 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-host-cni-netd\") pod \"bd007c61-4689-4179-95ad-44a9724e8e7e\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.924715 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-run-ovn\") pod \"bd007c61-4689-4179-95ad-44a9724e8e7e\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.924740 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-host-run-netns\") pod \"bd007c61-4689-4179-95ad-44a9724e8e7e\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.924776 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-var-lib-openvswitch\") pod \"bd007c61-4689-4179-95ad-44a9724e8e7e\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.924822 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/bd007c61-4689-4179-95ad-44a9724e8e7e-ovnkube-script-lib\") pod \"bd007c61-4689-4179-95ad-44a9724e8e7e\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.924857 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bd007c61-4689-4179-95ad-44a9724e8e7e-ovnkube-config\") pod \"bd007c61-4689-4179-95ad-44a9724e8e7e\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.924894 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-host-kubelet\") pod \"bd007c61-4689-4179-95ad-44a9724e8e7e\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.924913 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "bd007c61-4689-4179-95ad-44a9724e8e7e" (UID: "bd007c61-4689-4179-95ad-44a9724e8e7e"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.924924 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"bd007c61-4689-4179-95ad-44a9724e8e7e\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.924959 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "bd007c61-4689-4179-95ad-44a9724e8e7e" (UID: "bd007c61-4689-4179-95ad-44a9724e8e7e"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.924977 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-node-log\") pod \"bd007c61-4689-4179-95ad-44a9724e8e7e\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.924997 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-host-slash\") pod \"bd007c61-4689-4179-95ad-44a9724e8e7e\" (UID: \"bd007c61-4689-4179-95ad-44a9724e8e7e\") " Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.925275 4799 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.925289 4799 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.924995 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-log-socket" (OuterVolumeSpecName: "log-socket") pod "bd007c61-4689-4179-95ad-44a9724e8e7e" (UID: "bd007c61-4689-4179-95ad-44a9724e8e7e"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.925015 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "bd007c61-4689-4179-95ad-44a9724e8e7e" (UID: "bd007c61-4689-4179-95ad-44a9724e8e7e"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.925033 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "bd007c61-4689-4179-95ad-44a9724e8e7e" (UID: "bd007c61-4689-4179-95ad-44a9724e8e7e"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.925051 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "bd007c61-4689-4179-95ad-44a9724e8e7e" (UID: "bd007c61-4689-4179-95ad-44a9724e8e7e"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.925545 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd007c61-4689-4179-95ad-44a9724e8e7e-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "bd007c61-4689-4179-95ad-44a9724e8e7e" (UID: "bd007c61-4689-4179-95ad-44a9724e8e7e"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.925850 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-node-log" (OuterVolumeSpecName: "node-log") pod "bd007c61-4689-4179-95ad-44a9724e8e7e" (UID: "bd007c61-4689-4179-95ad-44a9724e8e7e"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.925867 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-host-slash" (OuterVolumeSpecName: "host-slash") pod "bd007c61-4689-4179-95ad-44a9724e8e7e" (UID: "bd007c61-4689-4179-95ad-44a9724e8e7e"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.926226 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "bd007c61-4689-4179-95ad-44a9724e8e7e" (UID: "bd007c61-4689-4179-95ad-44a9724e8e7e"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.927390 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "bd007c61-4689-4179-95ad-44a9724e8e7e" (UID: "bd007c61-4689-4179-95ad-44a9724e8e7e"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.927674 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "bd007c61-4689-4179-95ad-44a9724e8e7e" (UID: "bd007c61-4689-4179-95ad-44a9724e8e7e"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.928059 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "bd007c61-4689-4179-95ad-44a9724e8e7e" (UID: "bd007c61-4689-4179-95ad-44a9724e8e7e"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.928090 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd007c61-4689-4179-95ad-44a9724e8e7e-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "bd007c61-4689-4179-95ad-44a9724e8e7e" (UID: "bd007c61-4689-4179-95ad-44a9724e8e7e"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.928120 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "bd007c61-4689-4179-95ad-44a9724e8e7e" (UID: "bd007c61-4689-4179-95ad-44a9724e8e7e"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.928159 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "bd007c61-4689-4179-95ad-44a9724e8e7e" (UID: "bd007c61-4689-4179-95ad-44a9724e8e7e"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.928171 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd007c61-4689-4179-95ad-44a9724e8e7e-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "bd007c61-4689-4179-95ad-44a9724e8e7e" (UID: "bd007c61-4689-4179-95ad-44a9724e8e7e"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.931681 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd007c61-4689-4179-95ad-44a9724e8e7e-kube-api-access-jpd4j" (OuterVolumeSpecName: "kube-api-access-jpd4j") pod "bd007c61-4689-4179-95ad-44a9724e8e7e" (UID: "bd007c61-4689-4179-95ad-44a9724e8e7e"). InnerVolumeSpecName "kube-api-access-jpd4j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.931811 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd007c61-4689-4179-95ad-44a9724e8e7e-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "bd007c61-4689-4179-95ad-44a9724e8e7e" (UID: "bd007c61-4689-4179-95ad-44a9724e8e7e"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.941498 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "bd007c61-4689-4179-95ad-44a9724e8e7e" (UID: "bd007c61-4689-4179-95ad-44a9724e8e7e"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.947818 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-96nt6"] Nov 29 04:47:56 crc kubenswrapper[4799]: E1129 04:47:56.948085 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerName="kubecfg-setup" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.948106 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerName="kubecfg-setup" Nov 29 04:47:56 crc kubenswrapper[4799]: E1129 04:47:56.948116 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerName="northd" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.948123 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerName="northd" Nov 29 04:47:56 crc kubenswrapper[4799]: E1129 04:47:56.948129 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerName="nbdb" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.948135 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerName="nbdb" Nov 29 04:47:56 crc kubenswrapper[4799]: E1129 04:47:56.948148 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerName="ovnkube-controller" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.948154 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerName="ovnkube-controller" Nov 29 04:47:56 crc kubenswrapper[4799]: E1129 04:47:56.948164 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerName="ovn-acl-logging" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.948170 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerName="ovn-acl-logging" Nov 29 04:47:56 crc kubenswrapper[4799]: E1129 04:47:56.948179 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerName="ovn-controller" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.948185 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerName="ovn-controller" Nov 29 04:47:56 crc kubenswrapper[4799]: E1129 04:47:56.948193 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerName="sbdb" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.948200 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerName="sbdb" Nov 29 04:47:56 crc kubenswrapper[4799]: E1129 04:47:56.948209 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerName="ovnkube-controller" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.948215 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerName="ovnkube-controller" Nov 29 04:47:56 crc kubenswrapper[4799]: E1129 04:47:56.948223 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerName="ovnkube-controller" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.948229 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerName="ovnkube-controller" Nov 29 04:47:56 crc kubenswrapper[4799]: E1129 04:47:56.948238 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerName="kube-rbac-proxy-ovn-metrics" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.948243 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerName="kube-rbac-proxy-ovn-metrics" Nov 29 04:47:56 crc kubenswrapper[4799]: E1129 04:47:56.948252 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerName="kube-rbac-proxy-node" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.948259 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerName="kube-rbac-proxy-node" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.948378 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerName="kube-rbac-proxy-ovn-metrics" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.948389 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerName="ovn-acl-logging" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.948395 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerName="sbdb" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.948403 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerName="ovnkube-controller" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.948409 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerName="northd" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.948420 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerName="ovnkube-controller" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.948429 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerName="nbdb" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.948441 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerName="ovn-controller" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.948448 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerName="ovnkube-controller" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.948456 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerName="kube-rbac-proxy-node" Nov 29 04:47:56 crc kubenswrapper[4799]: E1129 04:47:56.948548 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerName="ovnkube-controller" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.948556 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerName="ovnkube-controller" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.948662 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd007c61-4689-4179-95ad-44a9724e8e7e" containerName="ovnkube-controller" Nov 29 04:47:56 crc kubenswrapper[4799]: I1129 04:47:56.950475 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.027142 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d2521887-e7e4-4e6d-b743-89cb6cc270a6-node-log\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.027189 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d2521887-e7e4-4e6d-b743-89cb6cc270a6-ovnkube-config\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.027218 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d2521887-e7e4-4e6d-b743-89cb6cc270a6-host-slash\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.027240 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d2521887-e7e4-4e6d-b743-89cb6cc270a6-run-openvswitch\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.027266 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqvlw\" (UniqueName: \"kubernetes.io/projected/d2521887-e7e4-4e6d-b743-89cb6cc270a6-kube-api-access-gqvlw\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.027427 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d2521887-e7e4-4e6d-b743-89cb6cc270a6-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.027561 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d2521887-e7e4-4e6d-b743-89cb6cc270a6-ovn-node-metrics-cert\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.027689 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d2521887-e7e4-4e6d-b743-89cb6cc270a6-run-systemd\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.027744 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d2521887-e7e4-4e6d-b743-89cb6cc270a6-run-ovn\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.027884 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d2521887-e7e4-4e6d-b743-89cb6cc270a6-etc-openvswitch\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.027950 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d2521887-e7e4-4e6d-b743-89cb6cc270a6-log-socket\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.028145 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d2521887-e7e4-4e6d-b743-89cb6cc270a6-systemd-units\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.028254 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d2521887-e7e4-4e6d-b743-89cb6cc270a6-host-cni-bin\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.028343 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d2521887-e7e4-4e6d-b743-89cb6cc270a6-host-kubelet\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.028505 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d2521887-e7e4-4e6d-b743-89cb6cc270a6-var-lib-openvswitch\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.028590 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d2521887-e7e4-4e6d-b743-89cb6cc270a6-host-cni-netd\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.028820 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d2521887-e7e4-4e6d-b743-89cb6cc270a6-host-run-ovn-kubernetes\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.028876 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d2521887-e7e4-4e6d-b743-89cb6cc270a6-ovnkube-script-lib\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.028944 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d2521887-e7e4-4e6d-b743-89cb6cc270a6-env-overrides\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.028994 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d2521887-e7e4-4e6d-b743-89cb6cc270a6-host-run-netns\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.029116 4799 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-host-kubelet\") on node \"crc\" DevicePath \"\"" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.029149 4799 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-node-log\") on node \"crc\" DevicePath \"\"" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.029173 4799 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-host-slash\") on node \"crc\" DevicePath \"\"" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.029192 4799 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-host-cni-bin\") on node \"crc\" DevicePath \"\"" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.029211 4799 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-log-socket\") on node \"crc\" DevicePath \"\"" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.029231 4799 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-run-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.029248 4799 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-systemd-units\") on node \"crc\" DevicePath \"\"" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.029268 4799 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.029288 4799 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bd007c61-4689-4179-95ad-44a9724e8e7e-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.029308 4799 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-run-systemd\") on node \"crc\" DevicePath \"\"" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.029325 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jpd4j\" (UniqueName: \"kubernetes.io/projected/bd007c61-4689-4179-95ad-44a9724e8e7e-kube-api-access-jpd4j\") on node \"crc\" DevicePath \"\"" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.029344 4799 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bd007c61-4689-4179-95ad-44a9724e8e7e-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.029366 4799 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-host-cni-netd\") on node \"crc\" DevicePath \"\"" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.029389 4799 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.029407 4799 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-host-run-netns\") on node \"crc\" DevicePath \"\"" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.029424 4799 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bd007c61-4689-4179-95ad-44a9724e8e7e-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.029443 4799 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/bd007c61-4689-4179-95ad-44a9724e8e7e-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.029461 4799 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bd007c61-4689-4179-95ad-44a9724e8e7e-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.130543 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d2521887-e7e4-4e6d-b743-89cb6cc270a6-host-slash\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.130590 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d2521887-e7e4-4e6d-b743-89cb6cc270a6-run-openvswitch\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.130638 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqvlw\" (UniqueName: \"kubernetes.io/projected/d2521887-e7e4-4e6d-b743-89cb6cc270a6-kube-api-access-gqvlw\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.130661 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d2521887-e7e4-4e6d-b743-89cb6cc270a6-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.130688 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d2521887-e7e4-4e6d-b743-89cb6cc270a6-ovn-node-metrics-cert\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.130715 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d2521887-e7e4-4e6d-b743-89cb6cc270a6-run-ovn\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.130732 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d2521887-e7e4-4e6d-b743-89cb6cc270a6-run-systemd\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.131463 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d2521887-e7e4-4e6d-b743-89cb6cc270a6-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.131486 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d2521887-e7e4-4e6d-b743-89cb6cc270a6-host-slash\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.131529 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d2521887-e7e4-4e6d-b743-89cb6cc270a6-run-openvswitch\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.131556 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d2521887-e7e4-4e6d-b743-89cb6cc270a6-run-ovn\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.131607 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d2521887-e7e4-4e6d-b743-89cb6cc270a6-run-systemd\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.131774 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d2521887-e7e4-4e6d-b743-89cb6cc270a6-etc-openvswitch\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.131839 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d2521887-e7e4-4e6d-b743-89cb6cc270a6-log-socket\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.131914 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d2521887-e7e4-4e6d-b743-89cb6cc270a6-systemd-units\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.131979 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d2521887-e7e4-4e6d-b743-89cb6cc270a6-host-cni-bin\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.132026 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d2521887-e7e4-4e6d-b743-89cb6cc270a6-log-socket\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.132035 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d2521887-e7e4-4e6d-b743-89cb6cc270a6-etc-openvswitch\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.132132 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d2521887-e7e4-4e6d-b743-89cb6cc270a6-host-cni-bin\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.132266 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d2521887-e7e4-4e6d-b743-89cb6cc270a6-host-kubelet\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.132062 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d2521887-e7e4-4e6d-b743-89cb6cc270a6-systemd-units\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.134866 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d2521887-e7e4-4e6d-b743-89cb6cc270a6-host-kubelet\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.134997 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d2521887-e7e4-4e6d-b743-89cb6cc270a6-var-lib-openvswitch\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.135156 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d2521887-e7e4-4e6d-b743-89cb6cc270a6-host-cni-netd\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.135258 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d2521887-e7e4-4e6d-b743-89cb6cc270a6-host-run-ovn-kubernetes\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.135339 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d2521887-e7e4-4e6d-b743-89cb6cc270a6-ovnkube-script-lib\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.135402 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d2521887-e7e4-4e6d-b743-89cb6cc270a6-env-overrides\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.135439 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d2521887-e7e4-4e6d-b743-89cb6cc270a6-host-run-netns\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.135516 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d2521887-e7e4-4e6d-b743-89cb6cc270a6-node-log\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.135576 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d2521887-e7e4-4e6d-b743-89cb6cc270a6-ovnkube-config\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.135956 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d2521887-e7e4-4e6d-b743-89cb6cc270a6-host-run-netns\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.136769 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d2521887-e7e4-4e6d-b743-89cb6cc270a6-ovnkube-script-lib\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.136891 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d2521887-e7e4-4e6d-b743-89cb6cc270a6-host-run-ovn-kubernetes\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.136895 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d2521887-e7e4-4e6d-b743-89cb6cc270a6-host-cni-netd\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.135065 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d2521887-e7e4-4e6d-b743-89cb6cc270a6-var-lib-openvswitch\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.136978 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d2521887-e7e4-4e6d-b743-89cb6cc270a6-node-log\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.137485 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d2521887-e7e4-4e6d-b743-89cb6cc270a6-env-overrides\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.138345 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d2521887-e7e4-4e6d-b743-89cb6cc270a6-ovnkube-config\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.140330 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d2521887-e7e4-4e6d-b743-89cb6cc270a6-ovn-node-metrics-cert\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.164415 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqvlw\" (UniqueName: \"kubernetes.io/projected/d2521887-e7e4-4e6d-b743-89cb6cc270a6-kube-api-access-gqvlw\") pod \"ovnkube-node-96nt6\" (UID: \"d2521887-e7e4-4e6d-b743-89cb6cc270a6\") " pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.266376 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.831550 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fht8l_bd007c61-4689-4179-95ad-44a9724e8e7e/ovn-acl-logging/0.log" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.833014 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fht8l_bd007c61-4689-4179-95ad-44a9724e8e7e/ovn-controller/0.log" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.833604 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" event={"ID":"bd007c61-4689-4179-95ad-44a9724e8e7e","Type":"ContainerDied","Data":"c250bd8866cb517877d111014844a268897f17a961f506bc7259d08e24036a0a"} Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.833717 4799 scope.go:117] "RemoveContainer" containerID="66dd6f66933db1fc65c8aa5ac9ce4117213e4a0a3f1260975e0d9a6e06366f3d" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.834020 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-fht8l" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.836741 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-qt2dl_3b6c9236-5f7d-46c5-b080-4667620d874e/kube-multus/1.log" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.840005 4799 generic.go:334] "Generic (PLEG): container finished" podID="d2521887-e7e4-4e6d-b743-89cb6cc270a6" containerID="6eca1f5acb2f834333119e589dbb5d1f00f701b68a1f870291c1a3616bd8a14f" exitCode=0 Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.840069 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" event={"ID":"d2521887-e7e4-4e6d-b743-89cb6cc270a6","Type":"ContainerDied","Data":"6eca1f5acb2f834333119e589dbb5d1f00f701b68a1f870291c1a3616bd8a14f"} Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.840107 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" event={"ID":"d2521887-e7e4-4e6d-b743-89cb6cc270a6","Type":"ContainerStarted","Data":"3fd232b02f1a7b24b456f100d2b8afe8bdc0109c09a136e776a8d86f7f059d9b"} Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.855502 4799 scope.go:117] "RemoveContainer" containerID="d135e42b8886f1f118abf8f4ce4cf2875ac4dc6c148bfc0677cef6b3651562c0" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.879509 4799 scope.go:117] "RemoveContainer" containerID="72802532a47b381ead2507e1ae31c49852f9b1d332f3c08fb1e3a972a128eead" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.900673 4799 scope.go:117] "RemoveContainer" containerID="0c81d94f098efec48c1ff8e172e476bf806ccec5777974eea9c0f3850e5fd40e" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.911348 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-fht8l"] Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.923089 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-fht8l"] Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.930144 4799 scope.go:117] "RemoveContainer" containerID="dac1c084955d874741d36709fa848c2c475a2d7108d4378b677bb13a88184105" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.957583 4799 scope.go:117] "RemoveContainer" containerID="28499112c33332c23be39a11cdf7fc4e1946d04524a3f7c534087179a6620121" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.972809 4799 scope.go:117] "RemoveContainer" containerID="e3d7f1ba1a408c54f0470fca40e307b70ba208dc7eeece599b684820e3f56c41" Nov 29 04:47:57 crc kubenswrapper[4799]: I1129 04:47:57.989240 4799 scope.go:117] "RemoveContainer" containerID="8e76d91c4f461a1912117cb8ed3a8087c5efaaf0a67455b721bfbeaad1c7dda8" Nov 29 04:47:58 crc kubenswrapper[4799]: I1129 04:47:58.008517 4799 scope.go:117] "RemoveContainer" containerID="c9cb824a0e0e687614c97abd450d60df33f14f8d8971db46594984cae42b4e64" Nov 29 04:47:58 crc kubenswrapper[4799]: I1129 04:47:58.671173 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd007c61-4689-4179-95ad-44a9724e8e7e" path="/var/lib/kubelet/pods/bd007c61-4689-4179-95ad-44a9724e8e7e/volumes" Nov 29 04:47:58 crc kubenswrapper[4799]: I1129 04:47:58.850799 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" event={"ID":"d2521887-e7e4-4e6d-b743-89cb6cc270a6","Type":"ContainerStarted","Data":"c99f1cc2bf36326f390b790ca5514e9c6291b5373085ca665cae730abb2861a2"} Nov 29 04:47:58 crc kubenswrapper[4799]: I1129 04:47:58.850855 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" event={"ID":"d2521887-e7e4-4e6d-b743-89cb6cc270a6","Type":"ContainerStarted","Data":"d31dfa225f7fecbef6aa7e241ae9386935bd62f2de1a8337115899a067485959"} Nov 29 04:47:58 crc kubenswrapper[4799]: I1129 04:47:58.850867 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" event={"ID":"d2521887-e7e4-4e6d-b743-89cb6cc270a6","Type":"ContainerStarted","Data":"2fdb0a95c2aeb28ee772d4cc6e009d04578101fcf9e4e8dfb16047ad8d61b529"} Nov 29 04:47:58 crc kubenswrapper[4799]: I1129 04:47:58.850878 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" event={"ID":"d2521887-e7e4-4e6d-b743-89cb6cc270a6","Type":"ContainerStarted","Data":"436c73ed478bc964359b05d677b18dcbbbff6dc76cb114f0fda464745fd7dbfe"} Nov 29 04:47:58 crc kubenswrapper[4799]: I1129 04:47:58.850888 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" event={"ID":"d2521887-e7e4-4e6d-b743-89cb6cc270a6","Type":"ContainerStarted","Data":"2e4efc5fd44ff93de99eec443c15cfcdce7af1f1b38b3358cb3a58225d2f87db"} Nov 29 04:47:58 crc kubenswrapper[4799]: I1129 04:47:58.850898 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" event={"ID":"d2521887-e7e4-4e6d-b743-89cb6cc270a6","Type":"ContainerStarted","Data":"c6418646900ca5f7eb90a905e75e6abf2dc92a863b37cd2a962382e20f099cde"} Nov 29 04:48:01 crc kubenswrapper[4799]: I1129 04:48:01.884605 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" event={"ID":"d2521887-e7e4-4e6d-b743-89cb6cc270a6","Type":"ContainerStarted","Data":"e6210d12d72ec60683bb7b0c934b81789b634067e56c1623b7f7cf75951dd893"} Nov 29 04:48:03 crc kubenswrapper[4799]: I1129 04:48:03.900873 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" event={"ID":"d2521887-e7e4-4e6d-b743-89cb6cc270a6","Type":"ContainerStarted","Data":"f4f89fd4d710844bb82fde24f7c37b0d19a511168b990c38ffe9be9bfcbc7e2b"} Nov 29 04:48:03 crc kubenswrapper[4799]: I1129 04:48:03.901400 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:48:03 crc kubenswrapper[4799]: I1129 04:48:03.901419 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:48:03 crc kubenswrapper[4799]: I1129 04:48:03.901442 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:48:03 crc kubenswrapper[4799]: I1129 04:48:03.932382 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:48:03 crc kubenswrapper[4799]: I1129 04:48:03.947276 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:48:03 crc kubenswrapper[4799]: I1129 04:48:03.950403 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" podStartSLOduration=7.950384421 podStartE2EDuration="7.950384421s" podCreationTimestamp="2025-11-29 04:47:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:48:03.94762159 +0000 UTC m=+559.590552000" watchObservedRunningTime="2025-11-29 04:48:03.950384421 +0000 UTC m=+559.593314821" Nov 29 04:48:07 crc kubenswrapper[4799]: I1129 04:48:07.498696 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 04:48:07 crc kubenswrapper[4799]: I1129 04:48:07.499379 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 04:48:07 crc kubenswrapper[4799]: I1129 04:48:07.659640 4799 scope.go:117] "RemoveContainer" containerID="d0e3c7d98a4a048822bc263c302eef3b130a53051289505324e512944f332b5d" Nov 29 04:48:08 crc kubenswrapper[4799]: I1129 04:48:08.935718 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-qt2dl_3b6c9236-5f7d-46c5-b080-4667620d874e/kube-multus/1.log" Nov 29 04:48:08 crc kubenswrapper[4799]: I1129 04:48:08.936264 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-qt2dl" event={"ID":"3b6c9236-5f7d-46c5-b080-4667620d874e","Type":"ContainerStarted","Data":"a105e26ef244581d215b60b37ae32546cecc10deb8ea69042f5f6cdc38e0c06e"} Nov 29 04:48:27 crc kubenswrapper[4799]: I1129 04:48:27.300118 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-96nt6" Nov 29 04:48:34 crc kubenswrapper[4799]: I1129 04:48:34.823583 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftbmdt"] Nov 29 04:48:34 crc kubenswrapper[4799]: I1129 04:48:34.826831 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftbmdt" Nov 29 04:48:34 crc kubenswrapper[4799]: I1129 04:48:34.829733 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 29 04:48:34 crc kubenswrapper[4799]: I1129 04:48:34.837723 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftbmdt"] Nov 29 04:48:34 crc kubenswrapper[4799]: I1129 04:48:34.896771 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/01a9ddbf-bda8-4f39-b22e-400f1a1505b0-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftbmdt\" (UID: \"01a9ddbf-bda8-4f39-b22e-400f1a1505b0\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftbmdt" Nov 29 04:48:34 crc kubenswrapper[4799]: I1129 04:48:34.896964 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8dmt\" (UniqueName: \"kubernetes.io/projected/01a9ddbf-bda8-4f39-b22e-400f1a1505b0-kube-api-access-g8dmt\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftbmdt\" (UID: \"01a9ddbf-bda8-4f39-b22e-400f1a1505b0\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftbmdt" Nov 29 04:48:34 crc kubenswrapper[4799]: I1129 04:48:34.897005 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/01a9ddbf-bda8-4f39-b22e-400f1a1505b0-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftbmdt\" (UID: \"01a9ddbf-bda8-4f39-b22e-400f1a1505b0\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftbmdt" Nov 29 04:48:34 crc kubenswrapper[4799]: I1129 04:48:34.998499 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8dmt\" (UniqueName: \"kubernetes.io/projected/01a9ddbf-bda8-4f39-b22e-400f1a1505b0-kube-api-access-g8dmt\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftbmdt\" (UID: \"01a9ddbf-bda8-4f39-b22e-400f1a1505b0\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftbmdt" Nov 29 04:48:34 crc kubenswrapper[4799]: I1129 04:48:34.998614 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/01a9ddbf-bda8-4f39-b22e-400f1a1505b0-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftbmdt\" (UID: \"01a9ddbf-bda8-4f39-b22e-400f1a1505b0\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftbmdt" Nov 29 04:48:34 crc kubenswrapper[4799]: I1129 04:48:34.998682 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/01a9ddbf-bda8-4f39-b22e-400f1a1505b0-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftbmdt\" (UID: \"01a9ddbf-bda8-4f39-b22e-400f1a1505b0\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftbmdt" Nov 29 04:48:34 crc kubenswrapper[4799]: I1129 04:48:34.999142 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/01a9ddbf-bda8-4f39-b22e-400f1a1505b0-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftbmdt\" (UID: \"01a9ddbf-bda8-4f39-b22e-400f1a1505b0\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftbmdt" Nov 29 04:48:34 crc kubenswrapper[4799]: I1129 04:48:34.999435 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/01a9ddbf-bda8-4f39-b22e-400f1a1505b0-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftbmdt\" (UID: \"01a9ddbf-bda8-4f39-b22e-400f1a1505b0\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftbmdt" Nov 29 04:48:35 crc kubenswrapper[4799]: I1129 04:48:35.018066 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8dmt\" (UniqueName: \"kubernetes.io/projected/01a9ddbf-bda8-4f39-b22e-400f1a1505b0-kube-api-access-g8dmt\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftbmdt\" (UID: \"01a9ddbf-bda8-4f39-b22e-400f1a1505b0\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftbmdt" Nov 29 04:48:35 crc kubenswrapper[4799]: I1129 04:48:35.188144 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftbmdt" Nov 29 04:48:35 crc kubenswrapper[4799]: I1129 04:48:35.397187 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftbmdt"] Nov 29 04:48:36 crc kubenswrapper[4799]: I1129 04:48:36.145650 4799 generic.go:334] "Generic (PLEG): container finished" podID="01a9ddbf-bda8-4f39-b22e-400f1a1505b0" containerID="fb7dc2e86a3373a6347bb5d5aae09271357b5d22f80dee2bfa66e047358d2265" exitCode=0 Nov 29 04:48:36 crc kubenswrapper[4799]: I1129 04:48:36.145934 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftbmdt" event={"ID":"01a9ddbf-bda8-4f39-b22e-400f1a1505b0","Type":"ContainerDied","Data":"fb7dc2e86a3373a6347bb5d5aae09271357b5d22f80dee2bfa66e047358d2265"} Nov 29 04:48:36 crc kubenswrapper[4799]: I1129 04:48:36.146417 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftbmdt" event={"ID":"01a9ddbf-bda8-4f39-b22e-400f1a1505b0","Type":"ContainerStarted","Data":"76ecb7c34547a9a79c025b49e65981571bdbc15fe4e96109b8d62ab4153db15c"} Nov 29 04:48:37 crc kubenswrapper[4799]: I1129 04:48:37.498715 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 04:48:37 crc kubenswrapper[4799]: I1129 04:48:37.498867 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 04:48:37 crc kubenswrapper[4799]: I1129 04:48:37.498954 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-26t88" Nov 29 04:48:37 crc kubenswrapper[4799]: I1129 04:48:37.500050 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2b1b7386c0bb5e909eb4074926c911c29f96ed91593e4e76b3689a1397aeb0ff"} pod="openshift-machine-config-operator/machine-config-daemon-26t88" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 29 04:48:37 crc kubenswrapper[4799]: I1129 04:48:37.500172 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" containerID="cri-o://2b1b7386c0bb5e909eb4074926c911c29f96ed91593e4e76b3689a1397aeb0ff" gracePeriod=600 Nov 29 04:48:38 crc kubenswrapper[4799]: I1129 04:48:38.164732 4799 generic.go:334] "Generic (PLEG): container finished" podID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerID="2b1b7386c0bb5e909eb4074926c911c29f96ed91593e4e76b3689a1397aeb0ff" exitCode=0 Nov 29 04:48:38 crc kubenswrapper[4799]: I1129 04:48:38.164779 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" event={"ID":"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8","Type":"ContainerDied","Data":"2b1b7386c0bb5e909eb4074926c911c29f96ed91593e4e76b3689a1397aeb0ff"} Nov 29 04:48:38 crc kubenswrapper[4799]: I1129 04:48:38.165290 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" event={"ID":"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8","Type":"ContainerStarted","Data":"99c5c08976d14d2fb8bfd60a883c0eaa07a2b4ec5ce6f9de4d8a5dd6e5c7212d"} Nov 29 04:48:38 crc kubenswrapper[4799]: I1129 04:48:38.165322 4799 scope.go:117] "RemoveContainer" containerID="0ae025ef68b62241dcb107a107b30c8479869f4f96294d65af8cc7c29fb998ff" Nov 29 04:48:38 crc kubenswrapper[4799]: I1129 04:48:38.167705 4799 generic.go:334] "Generic (PLEG): container finished" podID="01a9ddbf-bda8-4f39-b22e-400f1a1505b0" containerID="3007d4981785a36dcf13aceb899a034b86981f396118a66469b7a3086e1a0497" exitCode=0 Nov 29 04:48:38 crc kubenswrapper[4799]: I1129 04:48:38.167734 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftbmdt" event={"ID":"01a9ddbf-bda8-4f39-b22e-400f1a1505b0","Type":"ContainerDied","Data":"3007d4981785a36dcf13aceb899a034b86981f396118a66469b7a3086e1a0497"} Nov 29 04:48:39 crc kubenswrapper[4799]: I1129 04:48:39.184536 4799 generic.go:334] "Generic (PLEG): container finished" podID="01a9ddbf-bda8-4f39-b22e-400f1a1505b0" containerID="97db19786b53a8c65c603dc795a6e18b2004406330bd0a92f0201279ecd4dcaa" exitCode=0 Nov 29 04:48:39 crc kubenswrapper[4799]: I1129 04:48:39.184680 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftbmdt" event={"ID":"01a9ddbf-bda8-4f39-b22e-400f1a1505b0","Type":"ContainerDied","Data":"97db19786b53a8c65c603dc795a6e18b2004406330bd0a92f0201279ecd4dcaa"} Nov 29 04:48:40 crc kubenswrapper[4799]: I1129 04:48:40.466301 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftbmdt" Nov 29 04:48:40 crc kubenswrapper[4799]: I1129 04:48:40.489648 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/01a9ddbf-bda8-4f39-b22e-400f1a1505b0-bundle\") pod \"01a9ddbf-bda8-4f39-b22e-400f1a1505b0\" (UID: \"01a9ddbf-bda8-4f39-b22e-400f1a1505b0\") " Nov 29 04:48:40 crc kubenswrapper[4799]: I1129 04:48:40.489833 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8dmt\" (UniqueName: \"kubernetes.io/projected/01a9ddbf-bda8-4f39-b22e-400f1a1505b0-kube-api-access-g8dmt\") pod \"01a9ddbf-bda8-4f39-b22e-400f1a1505b0\" (UID: \"01a9ddbf-bda8-4f39-b22e-400f1a1505b0\") " Nov 29 04:48:40 crc kubenswrapper[4799]: I1129 04:48:40.489897 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/01a9ddbf-bda8-4f39-b22e-400f1a1505b0-util\") pod \"01a9ddbf-bda8-4f39-b22e-400f1a1505b0\" (UID: \"01a9ddbf-bda8-4f39-b22e-400f1a1505b0\") " Nov 29 04:48:40 crc kubenswrapper[4799]: I1129 04:48:40.491341 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01a9ddbf-bda8-4f39-b22e-400f1a1505b0-bundle" (OuterVolumeSpecName: "bundle") pod "01a9ddbf-bda8-4f39-b22e-400f1a1505b0" (UID: "01a9ddbf-bda8-4f39-b22e-400f1a1505b0"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:48:40 crc kubenswrapper[4799]: I1129 04:48:40.493201 4799 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/01a9ddbf-bda8-4f39-b22e-400f1a1505b0-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 04:48:40 crc kubenswrapper[4799]: I1129 04:48:40.498863 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01a9ddbf-bda8-4f39-b22e-400f1a1505b0-kube-api-access-g8dmt" (OuterVolumeSpecName: "kube-api-access-g8dmt") pod "01a9ddbf-bda8-4f39-b22e-400f1a1505b0" (UID: "01a9ddbf-bda8-4f39-b22e-400f1a1505b0"). InnerVolumeSpecName "kube-api-access-g8dmt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:48:40 crc kubenswrapper[4799]: I1129 04:48:40.510891 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01a9ddbf-bda8-4f39-b22e-400f1a1505b0-util" (OuterVolumeSpecName: "util") pod "01a9ddbf-bda8-4f39-b22e-400f1a1505b0" (UID: "01a9ddbf-bda8-4f39-b22e-400f1a1505b0"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:48:40 crc kubenswrapper[4799]: I1129 04:48:40.594582 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8dmt\" (UniqueName: \"kubernetes.io/projected/01a9ddbf-bda8-4f39-b22e-400f1a1505b0-kube-api-access-g8dmt\") on node \"crc\" DevicePath \"\"" Nov 29 04:48:40 crc kubenswrapper[4799]: I1129 04:48:40.594635 4799 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/01a9ddbf-bda8-4f39-b22e-400f1a1505b0-util\") on node \"crc\" DevicePath \"\"" Nov 29 04:48:41 crc kubenswrapper[4799]: I1129 04:48:41.208101 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftbmdt" event={"ID":"01a9ddbf-bda8-4f39-b22e-400f1a1505b0","Type":"ContainerDied","Data":"76ecb7c34547a9a79c025b49e65981571bdbc15fe4e96109b8d62ab4153db15c"} Nov 29 04:48:41 crc kubenswrapper[4799]: I1129 04:48:41.208490 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="76ecb7c34547a9a79c025b49e65981571bdbc15fe4e96109b8d62ab4153db15c" Nov 29 04:48:41 crc kubenswrapper[4799]: I1129 04:48:41.208591 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftbmdt" Nov 29 04:48:42 crc kubenswrapper[4799]: I1129 04:48:42.496986 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-7wvt6"] Nov 29 04:48:42 crc kubenswrapper[4799]: E1129 04:48:42.497270 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01a9ddbf-bda8-4f39-b22e-400f1a1505b0" containerName="extract" Nov 29 04:48:42 crc kubenswrapper[4799]: I1129 04:48:42.497283 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="01a9ddbf-bda8-4f39-b22e-400f1a1505b0" containerName="extract" Nov 29 04:48:42 crc kubenswrapper[4799]: E1129 04:48:42.497294 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01a9ddbf-bda8-4f39-b22e-400f1a1505b0" containerName="util" Nov 29 04:48:42 crc kubenswrapper[4799]: I1129 04:48:42.497300 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="01a9ddbf-bda8-4f39-b22e-400f1a1505b0" containerName="util" Nov 29 04:48:42 crc kubenswrapper[4799]: E1129 04:48:42.497319 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01a9ddbf-bda8-4f39-b22e-400f1a1505b0" containerName="pull" Nov 29 04:48:42 crc kubenswrapper[4799]: I1129 04:48:42.497325 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="01a9ddbf-bda8-4f39-b22e-400f1a1505b0" containerName="pull" Nov 29 04:48:42 crc kubenswrapper[4799]: I1129 04:48:42.497434 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="01a9ddbf-bda8-4f39-b22e-400f1a1505b0" containerName="extract" Nov 29 04:48:42 crc kubenswrapper[4799]: I1129 04:48:42.497912 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-7wvt6" Nov 29 04:48:42 crc kubenswrapper[4799]: I1129 04:48:42.501493 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Nov 29 04:48:42 crc kubenswrapper[4799]: I1129 04:48:42.501744 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Nov 29 04:48:42 crc kubenswrapper[4799]: I1129 04:48:42.502060 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-5r8qr" Nov 29 04:48:42 crc kubenswrapper[4799]: I1129 04:48:42.514959 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-7wvt6"] Nov 29 04:48:42 crc kubenswrapper[4799]: I1129 04:48:42.620890 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2wt9\" (UniqueName: \"kubernetes.io/projected/9924770d-6e30-4da4-959a-40b3eade5bf3-kube-api-access-m2wt9\") pod \"nmstate-operator-5b5b58f5c8-7wvt6\" (UID: \"9924770d-6e30-4da4-959a-40b3eade5bf3\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-7wvt6" Nov 29 04:48:42 crc kubenswrapper[4799]: I1129 04:48:42.723104 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2wt9\" (UniqueName: \"kubernetes.io/projected/9924770d-6e30-4da4-959a-40b3eade5bf3-kube-api-access-m2wt9\") pod \"nmstate-operator-5b5b58f5c8-7wvt6\" (UID: \"9924770d-6e30-4da4-959a-40b3eade5bf3\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-7wvt6" Nov 29 04:48:42 crc kubenswrapper[4799]: I1129 04:48:42.753111 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2wt9\" (UniqueName: \"kubernetes.io/projected/9924770d-6e30-4da4-959a-40b3eade5bf3-kube-api-access-m2wt9\") pod \"nmstate-operator-5b5b58f5c8-7wvt6\" (UID: \"9924770d-6e30-4da4-959a-40b3eade5bf3\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-7wvt6" Nov 29 04:48:42 crc kubenswrapper[4799]: I1129 04:48:42.812991 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-7wvt6" Nov 29 04:48:43 crc kubenswrapper[4799]: I1129 04:48:43.027229 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-7wvt6"] Nov 29 04:48:43 crc kubenswrapper[4799]: I1129 04:48:43.225039 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-7wvt6" event={"ID":"9924770d-6e30-4da4-959a-40b3eade5bf3","Type":"ContainerStarted","Data":"665eecbcb51d4f6f01ce23650d1d56c61cb5965b82059eefb5cc7d504d775b3b"} Nov 29 04:48:46 crc kubenswrapper[4799]: I1129 04:48:46.242838 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-7wvt6" event={"ID":"9924770d-6e30-4da4-959a-40b3eade5bf3","Type":"ContainerStarted","Data":"9a41db62167ae58c4823a6b919c15e1d4a30223cdc8a8761262863f846bec28e"} Nov 29 04:48:46 crc kubenswrapper[4799]: I1129 04:48:46.261532 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-7wvt6" podStartSLOduration=2.00154502 podStartE2EDuration="4.261417825s" podCreationTimestamp="2025-11-29 04:48:42 +0000 UTC" firstStartedPulling="2025-11-29 04:48:43.036472508 +0000 UTC m=+598.679402908" lastFinishedPulling="2025-11-29 04:48:45.296345313 +0000 UTC m=+600.939275713" observedRunningTime="2025-11-29 04:48:46.260504343 +0000 UTC m=+601.903434743" watchObservedRunningTime="2025-11-29 04:48:46.261417825 +0000 UTC m=+601.904348225" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.198943 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-2955x"] Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.200307 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-2955x" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.203190 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-f2hwd" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.206826 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2kxrf"] Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.207971 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2kxrf" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.210931 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.231109 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-7k5v8"] Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.232298 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-7k5v8" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.256589 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2kxrf"] Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.276056 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-2955x"] Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.288200 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/767c1d25-4a7d-49b3-8180-8f46793649be-ovs-socket\") pod \"nmstate-handler-7k5v8\" (UID: \"767c1d25-4a7d-49b3-8180-8f46793649be\") " pod="openshift-nmstate/nmstate-handler-7k5v8" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.288245 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfbdk\" (UniqueName: \"kubernetes.io/projected/767c1d25-4a7d-49b3-8180-8f46793649be-kube-api-access-tfbdk\") pod \"nmstate-handler-7k5v8\" (UID: \"767c1d25-4a7d-49b3-8180-8f46793649be\") " pod="openshift-nmstate/nmstate-handler-7k5v8" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.288270 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/767c1d25-4a7d-49b3-8180-8f46793649be-nmstate-lock\") pod \"nmstate-handler-7k5v8\" (UID: \"767c1d25-4a7d-49b3-8180-8f46793649be\") " pod="openshift-nmstate/nmstate-handler-7k5v8" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.288934 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/337961b2-38e7-4e32-831a-0c970535fecd-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-2kxrf\" (UID: \"337961b2-38e7-4e32-831a-0c970535fecd\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2kxrf" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.288958 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vnpv\" (UniqueName: \"kubernetes.io/projected/e96ac85b-718a-45dc-be96-b514bffa1431-kube-api-access-5vnpv\") pod \"nmstate-metrics-7f946cbc9-2955x\" (UID: \"e96ac85b-718a-45dc-be96-b514bffa1431\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-2955x" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.289017 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njsdk\" (UniqueName: \"kubernetes.io/projected/337961b2-38e7-4e32-831a-0c970535fecd-kube-api-access-njsdk\") pod \"nmstate-webhook-5f6d4c5ccb-2kxrf\" (UID: \"337961b2-38e7-4e32-831a-0c970535fecd\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2kxrf" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.289034 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/767c1d25-4a7d-49b3-8180-8f46793649be-dbus-socket\") pod \"nmstate-handler-7k5v8\" (UID: \"767c1d25-4a7d-49b3-8180-8f46793649be\") " pod="openshift-nmstate/nmstate-handler-7k5v8" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.361774 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-j5cvx"] Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.367079 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-j5cvx" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.369289 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-gj9rr" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.369656 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.369827 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.370484 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-j5cvx"] Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.390172 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/337961b2-38e7-4e32-831a-0c970535fecd-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-2kxrf\" (UID: \"337961b2-38e7-4e32-831a-0c970535fecd\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2kxrf" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.390224 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vnpv\" (UniqueName: \"kubernetes.io/projected/e96ac85b-718a-45dc-be96-b514bffa1431-kube-api-access-5vnpv\") pod \"nmstate-metrics-7f946cbc9-2955x\" (UID: \"e96ac85b-718a-45dc-be96-b514bffa1431\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-2955x" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.390249 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njsdk\" (UniqueName: \"kubernetes.io/projected/337961b2-38e7-4e32-831a-0c970535fecd-kube-api-access-njsdk\") pod \"nmstate-webhook-5f6d4c5ccb-2kxrf\" (UID: \"337961b2-38e7-4e32-831a-0c970535fecd\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2kxrf" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.390275 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/bb0e8eeb-0667-40e9-b814-33a66c0e78a0-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-j5cvx\" (UID: \"bb0e8eeb-0667-40e9-b814-33a66c0e78a0\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-j5cvx" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.390295 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/767c1d25-4a7d-49b3-8180-8f46793649be-dbus-socket\") pod \"nmstate-handler-7k5v8\" (UID: \"767c1d25-4a7d-49b3-8180-8f46793649be\") " pod="openshift-nmstate/nmstate-handler-7k5v8" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.390332 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/767c1d25-4a7d-49b3-8180-8f46793649be-ovs-socket\") pod \"nmstate-handler-7k5v8\" (UID: \"767c1d25-4a7d-49b3-8180-8f46793649be\") " pod="openshift-nmstate/nmstate-handler-7k5v8" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.390352 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfbdk\" (UniqueName: \"kubernetes.io/projected/767c1d25-4a7d-49b3-8180-8f46793649be-kube-api-access-tfbdk\") pod \"nmstate-handler-7k5v8\" (UID: \"767c1d25-4a7d-49b3-8180-8f46793649be\") " pod="openshift-nmstate/nmstate-handler-7k5v8" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.390371 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/767c1d25-4a7d-49b3-8180-8f46793649be-nmstate-lock\") pod \"nmstate-handler-7k5v8\" (UID: \"767c1d25-4a7d-49b3-8180-8f46793649be\") " pod="openshift-nmstate/nmstate-handler-7k5v8" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.390405 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/bb0e8eeb-0667-40e9-b814-33a66c0e78a0-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-j5cvx\" (UID: \"bb0e8eeb-0667-40e9-b814-33a66c0e78a0\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-j5cvx" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.390432 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kwdx\" (UniqueName: \"kubernetes.io/projected/bb0e8eeb-0667-40e9-b814-33a66c0e78a0-kube-api-access-7kwdx\") pod \"nmstate-console-plugin-7fbb5f6569-j5cvx\" (UID: \"bb0e8eeb-0667-40e9-b814-33a66c0e78a0\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-j5cvx" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.392087 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/767c1d25-4a7d-49b3-8180-8f46793649be-dbus-socket\") pod \"nmstate-handler-7k5v8\" (UID: \"767c1d25-4a7d-49b3-8180-8f46793649be\") " pod="openshift-nmstate/nmstate-handler-7k5v8" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.392163 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/767c1d25-4a7d-49b3-8180-8f46793649be-ovs-socket\") pod \"nmstate-handler-7k5v8\" (UID: \"767c1d25-4a7d-49b3-8180-8f46793649be\") " pod="openshift-nmstate/nmstate-handler-7k5v8" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.392173 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/767c1d25-4a7d-49b3-8180-8f46793649be-nmstate-lock\") pod \"nmstate-handler-7k5v8\" (UID: \"767c1d25-4a7d-49b3-8180-8f46793649be\") " pod="openshift-nmstate/nmstate-handler-7k5v8" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.399327 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/337961b2-38e7-4e32-831a-0c970535fecd-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-2kxrf\" (UID: \"337961b2-38e7-4e32-831a-0c970535fecd\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2kxrf" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.409358 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njsdk\" (UniqueName: \"kubernetes.io/projected/337961b2-38e7-4e32-831a-0c970535fecd-kube-api-access-njsdk\") pod \"nmstate-webhook-5f6d4c5ccb-2kxrf\" (UID: \"337961b2-38e7-4e32-831a-0c970535fecd\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2kxrf" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.415294 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vnpv\" (UniqueName: \"kubernetes.io/projected/e96ac85b-718a-45dc-be96-b514bffa1431-kube-api-access-5vnpv\") pod \"nmstate-metrics-7f946cbc9-2955x\" (UID: \"e96ac85b-718a-45dc-be96-b514bffa1431\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-2955x" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.415772 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfbdk\" (UniqueName: \"kubernetes.io/projected/767c1d25-4a7d-49b3-8180-8f46793649be-kube-api-access-tfbdk\") pod \"nmstate-handler-7k5v8\" (UID: \"767c1d25-4a7d-49b3-8180-8f46793649be\") " pod="openshift-nmstate/nmstate-handler-7k5v8" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.491374 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kwdx\" (UniqueName: \"kubernetes.io/projected/bb0e8eeb-0667-40e9-b814-33a66c0e78a0-kube-api-access-7kwdx\") pod \"nmstate-console-plugin-7fbb5f6569-j5cvx\" (UID: \"bb0e8eeb-0667-40e9-b814-33a66c0e78a0\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-j5cvx" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.491445 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/bb0e8eeb-0667-40e9-b814-33a66c0e78a0-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-j5cvx\" (UID: \"bb0e8eeb-0667-40e9-b814-33a66c0e78a0\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-j5cvx" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.491498 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/bb0e8eeb-0667-40e9-b814-33a66c0e78a0-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-j5cvx\" (UID: \"bb0e8eeb-0667-40e9-b814-33a66c0e78a0\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-j5cvx" Nov 29 04:48:47 crc kubenswrapper[4799]: E1129 04:48:47.492008 4799 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Nov 29 04:48:47 crc kubenswrapper[4799]: E1129 04:48:47.492161 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bb0e8eeb-0667-40e9-b814-33a66c0e78a0-plugin-serving-cert podName:bb0e8eeb-0667-40e9-b814-33a66c0e78a0 nodeName:}" failed. No retries permitted until 2025-11-29 04:48:47.992136756 +0000 UTC m=+603.635067156 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/bb0e8eeb-0667-40e9-b814-33a66c0e78a0-plugin-serving-cert") pod "nmstate-console-plugin-7fbb5f6569-j5cvx" (UID: "bb0e8eeb-0667-40e9-b814-33a66c0e78a0") : secret "plugin-serving-cert" not found Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.492524 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/bb0e8eeb-0667-40e9-b814-33a66c0e78a0-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-j5cvx\" (UID: \"bb0e8eeb-0667-40e9-b814-33a66c0e78a0\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-j5cvx" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.509245 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kwdx\" (UniqueName: \"kubernetes.io/projected/bb0e8eeb-0667-40e9-b814-33a66c0e78a0-kube-api-access-7kwdx\") pod \"nmstate-console-plugin-7fbb5f6569-j5cvx\" (UID: \"bb0e8eeb-0667-40e9-b814-33a66c0e78a0\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-j5cvx" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.527362 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-2955x" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.543554 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2kxrf" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.551350 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-6c77cc89ff-ss69s"] Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.552485 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6c77cc89ff-ss69s" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.553328 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-7k5v8" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.574249 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6c77cc89ff-ss69s"] Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.593765 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a65fb863-5c76-475e-b95e-c19340abf777-service-ca\") pod \"console-6c77cc89ff-ss69s\" (UID: \"a65fb863-5c76-475e-b95e-c19340abf777\") " pod="openshift-console/console-6c77cc89ff-ss69s" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.593860 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a65fb863-5c76-475e-b95e-c19340abf777-console-serving-cert\") pod \"console-6c77cc89ff-ss69s\" (UID: \"a65fb863-5c76-475e-b95e-c19340abf777\") " pod="openshift-console/console-6c77cc89ff-ss69s" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.593905 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a65fb863-5c76-475e-b95e-c19340abf777-oauth-serving-cert\") pod \"console-6c77cc89ff-ss69s\" (UID: \"a65fb863-5c76-475e-b95e-c19340abf777\") " pod="openshift-console/console-6c77cc89ff-ss69s" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.593928 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a65fb863-5c76-475e-b95e-c19340abf777-console-config\") pod \"console-6c77cc89ff-ss69s\" (UID: \"a65fb863-5c76-475e-b95e-c19340abf777\") " pod="openshift-console/console-6c77cc89ff-ss69s" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.593952 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a65fb863-5c76-475e-b95e-c19340abf777-trusted-ca-bundle\") pod \"console-6c77cc89ff-ss69s\" (UID: \"a65fb863-5c76-475e-b95e-c19340abf777\") " pod="openshift-console/console-6c77cc89ff-ss69s" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.593987 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqqgf\" (UniqueName: \"kubernetes.io/projected/a65fb863-5c76-475e-b95e-c19340abf777-kube-api-access-kqqgf\") pod \"console-6c77cc89ff-ss69s\" (UID: \"a65fb863-5c76-475e-b95e-c19340abf777\") " pod="openshift-console/console-6c77cc89ff-ss69s" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.594026 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a65fb863-5c76-475e-b95e-c19340abf777-console-oauth-config\") pod \"console-6c77cc89ff-ss69s\" (UID: \"a65fb863-5c76-475e-b95e-c19340abf777\") " pod="openshift-console/console-6c77cc89ff-ss69s" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.695293 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqqgf\" (UniqueName: \"kubernetes.io/projected/a65fb863-5c76-475e-b95e-c19340abf777-kube-api-access-kqqgf\") pod \"console-6c77cc89ff-ss69s\" (UID: \"a65fb863-5c76-475e-b95e-c19340abf777\") " pod="openshift-console/console-6c77cc89ff-ss69s" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.695377 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a65fb863-5c76-475e-b95e-c19340abf777-console-oauth-config\") pod \"console-6c77cc89ff-ss69s\" (UID: \"a65fb863-5c76-475e-b95e-c19340abf777\") " pod="openshift-console/console-6c77cc89ff-ss69s" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.695421 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a65fb863-5c76-475e-b95e-c19340abf777-service-ca\") pod \"console-6c77cc89ff-ss69s\" (UID: \"a65fb863-5c76-475e-b95e-c19340abf777\") " pod="openshift-console/console-6c77cc89ff-ss69s" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.695457 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a65fb863-5c76-475e-b95e-c19340abf777-console-serving-cert\") pod \"console-6c77cc89ff-ss69s\" (UID: \"a65fb863-5c76-475e-b95e-c19340abf777\") " pod="openshift-console/console-6c77cc89ff-ss69s" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.695502 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a65fb863-5c76-475e-b95e-c19340abf777-oauth-serving-cert\") pod \"console-6c77cc89ff-ss69s\" (UID: \"a65fb863-5c76-475e-b95e-c19340abf777\") " pod="openshift-console/console-6c77cc89ff-ss69s" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.695530 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a65fb863-5c76-475e-b95e-c19340abf777-console-config\") pod \"console-6c77cc89ff-ss69s\" (UID: \"a65fb863-5c76-475e-b95e-c19340abf777\") " pod="openshift-console/console-6c77cc89ff-ss69s" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.695561 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a65fb863-5c76-475e-b95e-c19340abf777-trusted-ca-bundle\") pod \"console-6c77cc89ff-ss69s\" (UID: \"a65fb863-5c76-475e-b95e-c19340abf777\") " pod="openshift-console/console-6c77cc89ff-ss69s" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.696695 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a65fb863-5c76-475e-b95e-c19340abf777-service-ca\") pod \"console-6c77cc89ff-ss69s\" (UID: \"a65fb863-5c76-475e-b95e-c19340abf777\") " pod="openshift-console/console-6c77cc89ff-ss69s" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.696945 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a65fb863-5c76-475e-b95e-c19340abf777-oauth-serving-cert\") pod \"console-6c77cc89ff-ss69s\" (UID: \"a65fb863-5c76-475e-b95e-c19340abf777\") " pod="openshift-console/console-6c77cc89ff-ss69s" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.697379 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a65fb863-5c76-475e-b95e-c19340abf777-trusted-ca-bundle\") pod \"console-6c77cc89ff-ss69s\" (UID: \"a65fb863-5c76-475e-b95e-c19340abf777\") " pod="openshift-console/console-6c77cc89ff-ss69s" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.698067 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a65fb863-5c76-475e-b95e-c19340abf777-console-config\") pod \"console-6c77cc89ff-ss69s\" (UID: \"a65fb863-5c76-475e-b95e-c19340abf777\") " pod="openshift-console/console-6c77cc89ff-ss69s" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.702246 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a65fb863-5c76-475e-b95e-c19340abf777-console-serving-cert\") pod \"console-6c77cc89ff-ss69s\" (UID: \"a65fb863-5c76-475e-b95e-c19340abf777\") " pod="openshift-console/console-6c77cc89ff-ss69s" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.703781 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a65fb863-5c76-475e-b95e-c19340abf777-console-oauth-config\") pod \"console-6c77cc89ff-ss69s\" (UID: \"a65fb863-5c76-475e-b95e-c19340abf777\") " pod="openshift-console/console-6c77cc89ff-ss69s" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.719675 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqqgf\" (UniqueName: \"kubernetes.io/projected/a65fb863-5c76-475e-b95e-c19340abf777-kube-api-access-kqqgf\") pod \"console-6c77cc89ff-ss69s\" (UID: \"a65fb863-5c76-475e-b95e-c19340abf777\") " pod="openshift-console/console-6c77cc89ff-ss69s" Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.904935 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2kxrf"] Nov 29 04:48:47 crc kubenswrapper[4799]: W1129 04:48:47.908664 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod337961b2_38e7_4e32_831a_0c970535fecd.slice/crio-3ec59ee0096760c48f9cdbc0bcf7aada84d356a1e838e5540f8f42b4a1e6245d WatchSource:0}: Error finding container 3ec59ee0096760c48f9cdbc0bcf7aada84d356a1e838e5540f8f42b4a1e6245d: Status 404 returned error can't find the container with id 3ec59ee0096760c48f9cdbc0bcf7aada84d356a1e838e5540f8f42b4a1e6245d Nov 29 04:48:47 crc kubenswrapper[4799]: I1129 04:48:47.987370 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6c77cc89ff-ss69s" Nov 29 04:48:48 crc kubenswrapper[4799]: I1129 04:48:48.001099 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/bb0e8eeb-0667-40e9-b814-33a66c0e78a0-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-j5cvx\" (UID: \"bb0e8eeb-0667-40e9-b814-33a66c0e78a0\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-j5cvx" Nov 29 04:48:48 crc kubenswrapper[4799]: I1129 04:48:48.004549 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/bb0e8eeb-0667-40e9-b814-33a66c0e78a0-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-j5cvx\" (UID: \"bb0e8eeb-0667-40e9-b814-33a66c0e78a0\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-j5cvx" Nov 29 04:48:48 crc kubenswrapper[4799]: I1129 04:48:48.069354 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-2955x"] Nov 29 04:48:48 crc kubenswrapper[4799]: I1129 04:48:48.178754 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6c77cc89ff-ss69s"] Nov 29 04:48:48 crc kubenswrapper[4799]: I1129 04:48:48.262362 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6c77cc89ff-ss69s" event={"ID":"a65fb863-5c76-475e-b95e-c19340abf777","Type":"ContainerStarted","Data":"237ef112f53ba674d47e98da21fa605c930f8c432215141742440e80412557cc"} Nov 29 04:48:48 crc kubenswrapper[4799]: I1129 04:48:48.263838 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-2955x" event={"ID":"e96ac85b-718a-45dc-be96-b514bffa1431","Type":"ContainerStarted","Data":"a4dee7f5b7bc4bc5f82ae9ddff243405e16ad99e379f50a5952fb5b56b5c8e23"} Nov 29 04:48:48 crc kubenswrapper[4799]: I1129 04:48:48.265472 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-7k5v8" event={"ID":"767c1d25-4a7d-49b3-8180-8f46793649be","Type":"ContainerStarted","Data":"8114e9fafc406be67e540728dab6a798db36c7575a28d6bd10d2e82eb6c3d0f8"} Nov 29 04:48:48 crc kubenswrapper[4799]: I1129 04:48:48.266399 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2kxrf" event={"ID":"337961b2-38e7-4e32-831a-0c970535fecd","Type":"ContainerStarted","Data":"3ec59ee0096760c48f9cdbc0bcf7aada84d356a1e838e5540f8f42b4a1e6245d"} Nov 29 04:48:48 crc kubenswrapper[4799]: I1129 04:48:48.281007 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-j5cvx" Nov 29 04:48:48 crc kubenswrapper[4799]: I1129 04:48:48.481055 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-j5cvx"] Nov 29 04:48:48 crc kubenswrapper[4799]: W1129 04:48:48.492389 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbb0e8eeb_0667_40e9_b814_33a66c0e78a0.slice/crio-0300e9f4fc2aa14e4192de4cee7c6fdeeab0cfc834787893046d471e796b290c WatchSource:0}: Error finding container 0300e9f4fc2aa14e4192de4cee7c6fdeeab0cfc834787893046d471e796b290c: Status 404 returned error can't find the container with id 0300e9f4fc2aa14e4192de4cee7c6fdeeab0cfc834787893046d471e796b290c Nov 29 04:48:49 crc kubenswrapper[4799]: I1129 04:48:49.284158 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6c77cc89ff-ss69s" event={"ID":"a65fb863-5c76-475e-b95e-c19340abf777","Type":"ContainerStarted","Data":"6269b5e3b700604300f6b3016dd721255265c2efd9c81653f07ba043582d4e38"} Nov 29 04:48:49 crc kubenswrapper[4799]: I1129 04:48:49.285853 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-j5cvx" event={"ID":"bb0e8eeb-0667-40e9-b814-33a66c0e78a0","Type":"ContainerStarted","Data":"0300e9f4fc2aa14e4192de4cee7c6fdeeab0cfc834787893046d471e796b290c"} Nov 29 04:48:49 crc kubenswrapper[4799]: I1129 04:48:49.303138 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-6c77cc89ff-ss69s" podStartSLOduration=2.303062859 podStartE2EDuration="2.303062859s" podCreationTimestamp="2025-11-29 04:48:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:48:49.301381317 +0000 UTC m=+604.944311717" watchObservedRunningTime="2025-11-29 04:48:49.303062859 +0000 UTC m=+604.945993259" Nov 29 04:48:51 crc kubenswrapper[4799]: I1129 04:48:51.311170 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2kxrf" event={"ID":"337961b2-38e7-4e32-831a-0c970535fecd","Type":"ContainerStarted","Data":"abfc8a12f1ade07d2e585318efa448e84b13b8f18fafa9b86763cf2fb0d71877"} Nov 29 04:48:51 crc kubenswrapper[4799]: I1129 04:48:51.312650 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2kxrf" Nov 29 04:48:51 crc kubenswrapper[4799]: I1129 04:48:51.316771 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-2955x" event={"ID":"e96ac85b-718a-45dc-be96-b514bffa1431","Type":"ContainerStarted","Data":"dc048f8fa8a7c8579daf35b4b603982a92c9aa1d7cdb7ea06419388d3fc43c69"} Nov 29 04:48:51 crc kubenswrapper[4799]: I1129 04:48:51.319631 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-7k5v8" event={"ID":"767c1d25-4a7d-49b3-8180-8f46793649be","Type":"ContainerStarted","Data":"b7c23fa1d142948c0a4270eb4a7dec6b59e59f7551b062500b894600226b4328"} Nov 29 04:48:51 crc kubenswrapper[4799]: I1129 04:48:51.319844 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-7k5v8" Nov 29 04:48:51 crc kubenswrapper[4799]: I1129 04:48:51.335195 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2kxrf" podStartSLOduration=1.974296522 podStartE2EDuration="4.335167571s" podCreationTimestamp="2025-11-29 04:48:47 +0000 UTC" firstStartedPulling="2025-11-29 04:48:47.911215541 +0000 UTC m=+603.554145961" lastFinishedPulling="2025-11-29 04:48:50.27208661 +0000 UTC m=+605.915017010" observedRunningTime="2025-11-29 04:48:51.330839743 +0000 UTC m=+606.973770153" watchObservedRunningTime="2025-11-29 04:48:51.335167571 +0000 UTC m=+606.978097971" Nov 29 04:48:51 crc kubenswrapper[4799]: I1129 04:48:51.359753 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-7k5v8" podStartSLOduration=1.804988542 podStartE2EDuration="4.359715435s" podCreationTimestamp="2025-11-29 04:48:47 +0000 UTC" firstStartedPulling="2025-11-29 04:48:47.684406003 +0000 UTC m=+603.327336403" lastFinishedPulling="2025-11-29 04:48:50.239132876 +0000 UTC m=+605.882063296" observedRunningTime="2025-11-29 04:48:51.353229293 +0000 UTC m=+606.996159713" watchObservedRunningTime="2025-11-29 04:48:51.359715435 +0000 UTC m=+607.002645835" Nov 29 04:48:52 crc kubenswrapper[4799]: I1129 04:48:52.327514 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-j5cvx" event={"ID":"bb0e8eeb-0667-40e9-b814-33a66c0e78a0","Type":"ContainerStarted","Data":"760dd84e4e9619544ad94960c1cb7d9ee96962cd469b76f5df4a45dc8e96434f"} Nov 29 04:48:52 crc kubenswrapper[4799]: I1129 04:48:52.346253 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-j5cvx" podStartSLOduration=2.641381074 podStartE2EDuration="5.346184311s" podCreationTimestamp="2025-11-29 04:48:47 +0000 UTC" firstStartedPulling="2025-11-29 04:48:48.49457042 +0000 UTC m=+604.137500820" lastFinishedPulling="2025-11-29 04:48:51.199373657 +0000 UTC m=+606.842304057" observedRunningTime="2025-11-29 04:48:52.342211342 +0000 UTC m=+607.985141742" watchObservedRunningTime="2025-11-29 04:48:52.346184311 +0000 UTC m=+607.989114711" Nov 29 04:48:53 crc kubenswrapper[4799]: I1129 04:48:53.338295 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-2955x" event={"ID":"e96ac85b-718a-45dc-be96-b514bffa1431","Type":"ContainerStarted","Data":"33579cc98194a993c4426e237a82ed47621c7293bbbbb6a9956060462b0ebb08"} Nov 29 04:48:57 crc kubenswrapper[4799]: I1129 04:48:57.590475 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-7k5v8" Nov 29 04:48:57 crc kubenswrapper[4799]: I1129 04:48:57.621679 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-2955x" podStartSLOduration=5.875346886 podStartE2EDuration="10.621646849s" podCreationTimestamp="2025-11-29 04:48:47 +0000 UTC" firstStartedPulling="2025-11-29 04:48:48.081397274 +0000 UTC m=+603.724327684" lastFinishedPulling="2025-11-29 04:48:52.827697247 +0000 UTC m=+608.470627647" observedRunningTime="2025-11-29 04:48:53.362396991 +0000 UTC m=+609.005327401" watchObservedRunningTime="2025-11-29 04:48:57.621646849 +0000 UTC m=+613.264577279" Nov 29 04:48:57 crc kubenswrapper[4799]: I1129 04:48:57.988426 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-6c77cc89ff-ss69s" Nov 29 04:48:57 crc kubenswrapper[4799]: I1129 04:48:57.988557 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-6c77cc89ff-ss69s" Nov 29 04:48:57 crc kubenswrapper[4799]: I1129 04:48:57.998667 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-6c77cc89ff-ss69s" Nov 29 04:48:58 crc kubenswrapper[4799]: I1129 04:48:58.382232 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-6c77cc89ff-ss69s" Nov 29 04:48:58 crc kubenswrapper[4799]: I1129 04:48:58.452617 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-2qb64"] Nov 29 04:49:07 crc kubenswrapper[4799]: I1129 04:49:07.551660 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2kxrf" Nov 29 04:49:21 crc kubenswrapper[4799]: I1129 04:49:21.514705 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83flptb"] Nov 29 04:49:21 crc kubenswrapper[4799]: I1129 04:49:21.516498 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83flptb" Nov 29 04:49:21 crc kubenswrapper[4799]: I1129 04:49:21.518930 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 29 04:49:21 crc kubenswrapper[4799]: I1129 04:49:21.540361 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83flptb"] Nov 29 04:49:21 crc kubenswrapper[4799]: I1129 04:49:21.626318 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c94962e4-e51f-4e30-9c0e-75a19dfa16d9-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83flptb\" (UID: \"c94962e4-e51f-4e30-9c0e-75a19dfa16d9\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83flptb" Nov 29 04:49:21 crc kubenswrapper[4799]: I1129 04:49:21.626368 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c94962e4-e51f-4e30-9c0e-75a19dfa16d9-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83flptb\" (UID: \"c94962e4-e51f-4e30-9c0e-75a19dfa16d9\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83flptb" Nov 29 04:49:21 crc kubenswrapper[4799]: I1129 04:49:21.626695 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slr2h\" (UniqueName: \"kubernetes.io/projected/c94962e4-e51f-4e30-9c0e-75a19dfa16d9-kube-api-access-slr2h\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83flptb\" (UID: \"c94962e4-e51f-4e30-9c0e-75a19dfa16d9\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83flptb" Nov 29 04:49:21 crc kubenswrapper[4799]: I1129 04:49:21.728454 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slr2h\" (UniqueName: \"kubernetes.io/projected/c94962e4-e51f-4e30-9c0e-75a19dfa16d9-kube-api-access-slr2h\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83flptb\" (UID: \"c94962e4-e51f-4e30-9c0e-75a19dfa16d9\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83flptb" Nov 29 04:49:21 crc kubenswrapper[4799]: I1129 04:49:21.728569 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c94962e4-e51f-4e30-9c0e-75a19dfa16d9-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83flptb\" (UID: \"c94962e4-e51f-4e30-9c0e-75a19dfa16d9\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83flptb" Nov 29 04:49:21 crc kubenswrapper[4799]: I1129 04:49:21.728617 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c94962e4-e51f-4e30-9c0e-75a19dfa16d9-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83flptb\" (UID: \"c94962e4-e51f-4e30-9c0e-75a19dfa16d9\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83flptb" Nov 29 04:49:21 crc kubenswrapper[4799]: I1129 04:49:21.729127 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c94962e4-e51f-4e30-9c0e-75a19dfa16d9-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83flptb\" (UID: \"c94962e4-e51f-4e30-9c0e-75a19dfa16d9\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83flptb" Nov 29 04:49:21 crc kubenswrapper[4799]: I1129 04:49:21.729197 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c94962e4-e51f-4e30-9c0e-75a19dfa16d9-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83flptb\" (UID: \"c94962e4-e51f-4e30-9c0e-75a19dfa16d9\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83flptb" Nov 29 04:49:21 crc kubenswrapper[4799]: I1129 04:49:21.760681 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slr2h\" (UniqueName: \"kubernetes.io/projected/c94962e4-e51f-4e30-9c0e-75a19dfa16d9-kube-api-access-slr2h\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83flptb\" (UID: \"c94962e4-e51f-4e30-9c0e-75a19dfa16d9\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83flptb" Nov 29 04:49:21 crc kubenswrapper[4799]: I1129 04:49:21.840270 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83flptb" Nov 29 04:49:22 crc kubenswrapper[4799]: I1129 04:49:22.248050 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83flptb"] Nov 29 04:49:22 crc kubenswrapper[4799]: I1129 04:49:22.551822 4799 generic.go:334] "Generic (PLEG): container finished" podID="c94962e4-e51f-4e30-9c0e-75a19dfa16d9" containerID="fc553618006ced9a5cf47293ba3b5352255edde682dff6949f238d9434853cdb" exitCode=0 Nov 29 04:49:22 crc kubenswrapper[4799]: I1129 04:49:22.551990 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83flptb" event={"ID":"c94962e4-e51f-4e30-9c0e-75a19dfa16d9","Type":"ContainerDied","Data":"fc553618006ced9a5cf47293ba3b5352255edde682dff6949f238d9434853cdb"} Nov 29 04:49:22 crc kubenswrapper[4799]: I1129 04:49:22.552277 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83flptb" event={"ID":"c94962e4-e51f-4e30-9c0e-75a19dfa16d9","Type":"ContainerStarted","Data":"a524e5587d2720a2403d9363f0b6df157a0e5b68a79785838b54c4fe2cc625d8"} Nov 29 04:49:23 crc kubenswrapper[4799]: I1129 04:49:23.518677 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-2qb64" podUID="b6794cf7-9161-499d-ba18-74b95caaf3ae" containerName="console" containerID="cri-o://de4da2900ad12d1a9926d491dac701179128832d39109e72fcb7ee0cad10ff84" gracePeriod=15 Nov 29 04:49:23 crc kubenswrapper[4799]: I1129 04:49:23.922679 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-2qb64_b6794cf7-9161-499d-ba18-74b95caaf3ae/console/0.log" Nov 29 04:49:23 crc kubenswrapper[4799]: I1129 04:49:23.922774 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-2qb64" Nov 29 04:49:23 crc kubenswrapper[4799]: I1129 04:49:23.967424 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rj74b\" (UniqueName: \"kubernetes.io/projected/b6794cf7-9161-499d-ba18-74b95caaf3ae-kube-api-access-rj74b\") pod \"b6794cf7-9161-499d-ba18-74b95caaf3ae\" (UID: \"b6794cf7-9161-499d-ba18-74b95caaf3ae\") " Nov 29 04:49:23 crc kubenswrapper[4799]: I1129 04:49:23.967483 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b6794cf7-9161-499d-ba18-74b95caaf3ae-console-serving-cert\") pod \"b6794cf7-9161-499d-ba18-74b95caaf3ae\" (UID: \"b6794cf7-9161-499d-ba18-74b95caaf3ae\") " Nov 29 04:49:23 crc kubenswrapper[4799]: I1129 04:49:23.967525 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b6794cf7-9161-499d-ba18-74b95caaf3ae-console-oauth-config\") pod \"b6794cf7-9161-499d-ba18-74b95caaf3ae\" (UID: \"b6794cf7-9161-499d-ba18-74b95caaf3ae\") " Nov 29 04:49:23 crc kubenswrapper[4799]: I1129 04:49:23.967605 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b6794cf7-9161-499d-ba18-74b95caaf3ae-trusted-ca-bundle\") pod \"b6794cf7-9161-499d-ba18-74b95caaf3ae\" (UID: \"b6794cf7-9161-499d-ba18-74b95caaf3ae\") " Nov 29 04:49:23 crc kubenswrapper[4799]: I1129 04:49:23.967634 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b6794cf7-9161-499d-ba18-74b95caaf3ae-oauth-serving-cert\") pod \"b6794cf7-9161-499d-ba18-74b95caaf3ae\" (UID: \"b6794cf7-9161-499d-ba18-74b95caaf3ae\") " Nov 29 04:49:23 crc kubenswrapper[4799]: I1129 04:49:23.967655 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b6794cf7-9161-499d-ba18-74b95caaf3ae-console-config\") pod \"b6794cf7-9161-499d-ba18-74b95caaf3ae\" (UID: \"b6794cf7-9161-499d-ba18-74b95caaf3ae\") " Nov 29 04:49:23 crc kubenswrapper[4799]: I1129 04:49:23.967755 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b6794cf7-9161-499d-ba18-74b95caaf3ae-service-ca\") pod \"b6794cf7-9161-499d-ba18-74b95caaf3ae\" (UID: \"b6794cf7-9161-499d-ba18-74b95caaf3ae\") " Nov 29 04:49:23 crc kubenswrapper[4799]: I1129 04:49:23.968641 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6794cf7-9161-499d-ba18-74b95caaf3ae-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "b6794cf7-9161-499d-ba18-74b95caaf3ae" (UID: "b6794cf7-9161-499d-ba18-74b95caaf3ae"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:49:23 crc kubenswrapper[4799]: I1129 04:49:23.968672 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6794cf7-9161-499d-ba18-74b95caaf3ae-service-ca" (OuterVolumeSpecName: "service-ca") pod "b6794cf7-9161-499d-ba18-74b95caaf3ae" (UID: "b6794cf7-9161-499d-ba18-74b95caaf3ae"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:49:23 crc kubenswrapper[4799]: I1129 04:49:23.968651 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6794cf7-9161-499d-ba18-74b95caaf3ae-console-config" (OuterVolumeSpecName: "console-config") pod "b6794cf7-9161-499d-ba18-74b95caaf3ae" (UID: "b6794cf7-9161-499d-ba18-74b95caaf3ae"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:49:23 crc kubenswrapper[4799]: I1129 04:49:23.968663 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6794cf7-9161-499d-ba18-74b95caaf3ae-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "b6794cf7-9161-499d-ba18-74b95caaf3ae" (UID: "b6794cf7-9161-499d-ba18-74b95caaf3ae"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:49:23 crc kubenswrapper[4799]: I1129 04:49:23.968973 4799 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b6794cf7-9161-499d-ba18-74b95caaf3ae-service-ca\") on node \"crc\" DevicePath \"\"" Nov 29 04:49:23 crc kubenswrapper[4799]: I1129 04:49:23.969281 4799 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b6794cf7-9161-499d-ba18-74b95caaf3ae-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 04:49:23 crc kubenswrapper[4799]: I1129 04:49:23.969293 4799 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b6794cf7-9161-499d-ba18-74b95caaf3ae-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 04:49:23 crc kubenswrapper[4799]: I1129 04:49:23.969301 4799 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b6794cf7-9161-499d-ba18-74b95caaf3ae-console-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:49:23 crc kubenswrapper[4799]: I1129 04:49:23.976442 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6794cf7-9161-499d-ba18-74b95caaf3ae-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "b6794cf7-9161-499d-ba18-74b95caaf3ae" (UID: "b6794cf7-9161-499d-ba18-74b95caaf3ae"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:49:23 crc kubenswrapper[4799]: I1129 04:49:23.976856 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6794cf7-9161-499d-ba18-74b95caaf3ae-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "b6794cf7-9161-499d-ba18-74b95caaf3ae" (UID: "b6794cf7-9161-499d-ba18-74b95caaf3ae"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:49:23 crc kubenswrapper[4799]: I1129 04:49:23.976933 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6794cf7-9161-499d-ba18-74b95caaf3ae-kube-api-access-rj74b" (OuterVolumeSpecName: "kube-api-access-rj74b") pod "b6794cf7-9161-499d-ba18-74b95caaf3ae" (UID: "b6794cf7-9161-499d-ba18-74b95caaf3ae"). InnerVolumeSpecName "kube-api-access-rj74b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:49:24 crc kubenswrapper[4799]: I1129 04:49:24.070451 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rj74b\" (UniqueName: \"kubernetes.io/projected/b6794cf7-9161-499d-ba18-74b95caaf3ae-kube-api-access-rj74b\") on node \"crc\" DevicePath \"\"" Nov 29 04:49:24 crc kubenswrapper[4799]: I1129 04:49:24.070909 4799 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b6794cf7-9161-499d-ba18-74b95caaf3ae-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 04:49:24 crc kubenswrapper[4799]: I1129 04:49:24.071099 4799 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b6794cf7-9161-499d-ba18-74b95caaf3ae-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:49:24 crc kubenswrapper[4799]: I1129 04:49:24.572819 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-2qb64_b6794cf7-9161-499d-ba18-74b95caaf3ae/console/0.log" Nov 29 04:49:24 crc kubenswrapper[4799]: I1129 04:49:24.572909 4799 generic.go:334] "Generic (PLEG): container finished" podID="b6794cf7-9161-499d-ba18-74b95caaf3ae" containerID="de4da2900ad12d1a9926d491dac701179128832d39109e72fcb7ee0cad10ff84" exitCode=2 Nov 29 04:49:24 crc kubenswrapper[4799]: I1129 04:49:24.573024 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-2qb64" event={"ID":"b6794cf7-9161-499d-ba18-74b95caaf3ae","Type":"ContainerDied","Data":"de4da2900ad12d1a9926d491dac701179128832d39109e72fcb7ee0cad10ff84"} Nov 29 04:49:24 crc kubenswrapper[4799]: I1129 04:49:24.573074 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-2qb64" event={"ID":"b6794cf7-9161-499d-ba18-74b95caaf3ae","Type":"ContainerDied","Data":"714279d9206d684d361016c0a993451532ec8a99d097a25966400cbd7b1d7d7d"} Nov 29 04:49:24 crc kubenswrapper[4799]: I1129 04:49:24.573087 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-2qb64" Nov 29 04:49:24 crc kubenswrapper[4799]: I1129 04:49:24.573109 4799 scope.go:117] "RemoveContainer" containerID="de4da2900ad12d1a9926d491dac701179128832d39109e72fcb7ee0cad10ff84" Nov 29 04:49:24 crc kubenswrapper[4799]: I1129 04:49:24.576338 4799 generic.go:334] "Generic (PLEG): container finished" podID="c94962e4-e51f-4e30-9c0e-75a19dfa16d9" containerID="6798da1aeded0e04f068640c3e21a5044aeff728cf26ce53ff7fe40dc2f78dcc" exitCode=0 Nov 29 04:49:24 crc kubenswrapper[4799]: I1129 04:49:24.576359 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83flptb" event={"ID":"c94962e4-e51f-4e30-9c0e-75a19dfa16d9","Type":"ContainerDied","Data":"6798da1aeded0e04f068640c3e21a5044aeff728cf26ce53ff7fe40dc2f78dcc"} Nov 29 04:49:24 crc kubenswrapper[4799]: I1129 04:49:24.619237 4799 scope.go:117] "RemoveContainer" containerID="de4da2900ad12d1a9926d491dac701179128832d39109e72fcb7ee0cad10ff84" Nov 29 04:49:24 crc kubenswrapper[4799]: E1129 04:49:24.620512 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de4da2900ad12d1a9926d491dac701179128832d39109e72fcb7ee0cad10ff84\": container with ID starting with de4da2900ad12d1a9926d491dac701179128832d39109e72fcb7ee0cad10ff84 not found: ID does not exist" containerID="de4da2900ad12d1a9926d491dac701179128832d39109e72fcb7ee0cad10ff84" Nov 29 04:49:24 crc kubenswrapper[4799]: I1129 04:49:24.620552 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de4da2900ad12d1a9926d491dac701179128832d39109e72fcb7ee0cad10ff84"} err="failed to get container status \"de4da2900ad12d1a9926d491dac701179128832d39109e72fcb7ee0cad10ff84\": rpc error: code = NotFound desc = could not find container \"de4da2900ad12d1a9926d491dac701179128832d39109e72fcb7ee0cad10ff84\": container with ID starting with de4da2900ad12d1a9926d491dac701179128832d39109e72fcb7ee0cad10ff84 not found: ID does not exist" Nov 29 04:49:24 crc kubenswrapper[4799]: I1129 04:49:24.648555 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-2qb64"] Nov 29 04:49:24 crc kubenswrapper[4799]: I1129 04:49:24.655402 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-2qb64"] Nov 29 04:49:24 crc kubenswrapper[4799]: I1129 04:49:24.668992 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6794cf7-9161-499d-ba18-74b95caaf3ae" path="/var/lib/kubelet/pods/b6794cf7-9161-499d-ba18-74b95caaf3ae/volumes" Nov 29 04:49:25 crc kubenswrapper[4799]: I1129 04:49:25.588178 4799 generic.go:334] "Generic (PLEG): container finished" podID="c94962e4-e51f-4e30-9c0e-75a19dfa16d9" containerID="ac1ff80e8d2cedc65a258a66bdd5776a6745e6d67d0ccc12ff8c4e7e7bd831cd" exitCode=0 Nov 29 04:49:25 crc kubenswrapper[4799]: I1129 04:49:25.588232 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83flptb" event={"ID":"c94962e4-e51f-4e30-9c0e-75a19dfa16d9","Type":"ContainerDied","Data":"ac1ff80e8d2cedc65a258a66bdd5776a6745e6d67d0ccc12ff8c4e7e7bd831cd"} Nov 29 04:49:26 crc kubenswrapper[4799]: I1129 04:49:26.906912 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83flptb" Nov 29 04:49:26 crc kubenswrapper[4799]: I1129 04:49:26.917074 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-slr2h\" (UniqueName: \"kubernetes.io/projected/c94962e4-e51f-4e30-9c0e-75a19dfa16d9-kube-api-access-slr2h\") pod \"c94962e4-e51f-4e30-9c0e-75a19dfa16d9\" (UID: \"c94962e4-e51f-4e30-9c0e-75a19dfa16d9\") " Nov 29 04:49:26 crc kubenswrapper[4799]: I1129 04:49:26.917508 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c94962e4-e51f-4e30-9c0e-75a19dfa16d9-bundle\") pod \"c94962e4-e51f-4e30-9c0e-75a19dfa16d9\" (UID: \"c94962e4-e51f-4e30-9c0e-75a19dfa16d9\") " Nov 29 04:49:26 crc kubenswrapper[4799]: I1129 04:49:26.917609 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c94962e4-e51f-4e30-9c0e-75a19dfa16d9-util\") pod \"c94962e4-e51f-4e30-9c0e-75a19dfa16d9\" (UID: \"c94962e4-e51f-4e30-9c0e-75a19dfa16d9\") " Nov 29 04:49:26 crc kubenswrapper[4799]: I1129 04:49:26.920371 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c94962e4-e51f-4e30-9c0e-75a19dfa16d9-bundle" (OuterVolumeSpecName: "bundle") pod "c94962e4-e51f-4e30-9c0e-75a19dfa16d9" (UID: "c94962e4-e51f-4e30-9c0e-75a19dfa16d9"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:49:26 crc kubenswrapper[4799]: I1129 04:49:26.946684 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c94962e4-e51f-4e30-9c0e-75a19dfa16d9-util" (OuterVolumeSpecName: "util") pod "c94962e4-e51f-4e30-9c0e-75a19dfa16d9" (UID: "c94962e4-e51f-4e30-9c0e-75a19dfa16d9"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:49:26 crc kubenswrapper[4799]: I1129 04:49:26.953984 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c94962e4-e51f-4e30-9c0e-75a19dfa16d9-kube-api-access-slr2h" (OuterVolumeSpecName: "kube-api-access-slr2h") pod "c94962e4-e51f-4e30-9c0e-75a19dfa16d9" (UID: "c94962e4-e51f-4e30-9c0e-75a19dfa16d9"). InnerVolumeSpecName "kube-api-access-slr2h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:49:27 crc kubenswrapper[4799]: I1129 04:49:27.021939 4799 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c94962e4-e51f-4e30-9c0e-75a19dfa16d9-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 04:49:27 crc kubenswrapper[4799]: I1129 04:49:27.021977 4799 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c94962e4-e51f-4e30-9c0e-75a19dfa16d9-util\") on node \"crc\" DevicePath \"\"" Nov 29 04:49:27 crc kubenswrapper[4799]: I1129 04:49:27.021991 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-slr2h\" (UniqueName: \"kubernetes.io/projected/c94962e4-e51f-4e30-9c0e-75a19dfa16d9-kube-api-access-slr2h\") on node \"crc\" DevicePath \"\"" Nov 29 04:49:27 crc kubenswrapper[4799]: I1129 04:49:27.606706 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83flptb" event={"ID":"c94962e4-e51f-4e30-9c0e-75a19dfa16d9","Type":"ContainerDied","Data":"a524e5587d2720a2403d9363f0b6df157a0e5b68a79785838b54c4fe2cc625d8"} Nov 29 04:49:27 crc kubenswrapper[4799]: I1129 04:49:27.606761 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a524e5587d2720a2403d9363f0b6df157a0e5b68a79785838b54c4fe2cc625d8" Nov 29 04:49:27 crc kubenswrapper[4799]: I1129 04:49:27.606895 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83flptb" Nov 29 04:49:36 crc kubenswrapper[4799]: I1129 04:49:36.774554 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-6887d9747b-t425t"] Nov 29 04:49:36 crc kubenswrapper[4799]: E1129 04:49:36.775383 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6794cf7-9161-499d-ba18-74b95caaf3ae" containerName="console" Nov 29 04:49:36 crc kubenswrapper[4799]: I1129 04:49:36.775396 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6794cf7-9161-499d-ba18-74b95caaf3ae" containerName="console" Nov 29 04:49:36 crc kubenswrapper[4799]: E1129 04:49:36.775412 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c94962e4-e51f-4e30-9c0e-75a19dfa16d9" containerName="util" Nov 29 04:49:36 crc kubenswrapper[4799]: I1129 04:49:36.775418 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="c94962e4-e51f-4e30-9c0e-75a19dfa16d9" containerName="util" Nov 29 04:49:36 crc kubenswrapper[4799]: E1129 04:49:36.775429 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c94962e4-e51f-4e30-9c0e-75a19dfa16d9" containerName="pull" Nov 29 04:49:36 crc kubenswrapper[4799]: I1129 04:49:36.775435 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="c94962e4-e51f-4e30-9c0e-75a19dfa16d9" containerName="pull" Nov 29 04:49:36 crc kubenswrapper[4799]: E1129 04:49:36.775445 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c94962e4-e51f-4e30-9c0e-75a19dfa16d9" containerName="extract" Nov 29 04:49:36 crc kubenswrapper[4799]: I1129 04:49:36.775452 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="c94962e4-e51f-4e30-9c0e-75a19dfa16d9" containerName="extract" Nov 29 04:49:36 crc kubenswrapper[4799]: I1129 04:49:36.775546 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6794cf7-9161-499d-ba18-74b95caaf3ae" containerName="console" Nov 29 04:49:36 crc kubenswrapper[4799]: I1129 04:49:36.775561 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="c94962e4-e51f-4e30-9c0e-75a19dfa16d9" containerName="extract" Nov 29 04:49:36 crc kubenswrapper[4799]: I1129 04:49:36.775978 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6887d9747b-t425t" Nov 29 04:49:36 crc kubenswrapper[4799]: I1129 04:49:36.778036 4799 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 29 04:49:36 crc kubenswrapper[4799]: I1129 04:49:36.778545 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 29 04:49:36 crc kubenswrapper[4799]: I1129 04:49:36.778640 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 29 04:49:36 crc kubenswrapper[4799]: I1129 04:49:36.778565 4799 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-j7kc9" Nov 29 04:49:36 crc kubenswrapper[4799]: I1129 04:49:36.780124 4799 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Nov 29 04:49:36 crc kubenswrapper[4799]: I1129 04:49:36.805407 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6887d9747b-t425t"] Nov 29 04:49:36 crc kubenswrapper[4799]: I1129 04:49:36.959489 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pv2p7\" (UniqueName: \"kubernetes.io/projected/cabb2055-e31c-4a75-a9f9-d403684efeec-kube-api-access-pv2p7\") pod \"metallb-operator-controller-manager-6887d9747b-t425t\" (UID: \"cabb2055-e31c-4a75-a9f9-d403684efeec\") " pod="metallb-system/metallb-operator-controller-manager-6887d9747b-t425t" Nov 29 04:49:36 crc kubenswrapper[4799]: I1129 04:49:36.959575 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cabb2055-e31c-4a75-a9f9-d403684efeec-webhook-cert\") pod \"metallb-operator-controller-manager-6887d9747b-t425t\" (UID: \"cabb2055-e31c-4a75-a9f9-d403684efeec\") " pod="metallb-system/metallb-operator-controller-manager-6887d9747b-t425t" Nov 29 04:49:36 crc kubenswrapper[4799]: I1129 04:49:36.959659 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cabb2055-e31c-4a75-a9f9-d403684efeec-apiservice-cert\") pod \"metallb-operator-controller-manager-6887d9747b-t425t\" (UID: \"cabb2055-e31c-4a75-a9f9-d403684efeec\") " pod="metallb-system/metallb-operator-controller-manager-6887d9747b-t425t" Nov 29 04:49:37 crc kubenswrapper[4799]: I1129 04:49:37.061349 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cabb2055-e31c-4a75-a9f9-d403684efeec-apiservice-cert\") pod \"metallb-operator-controller-manager-6887d9747b-t425t\" (UID: \"cabb2055-e31c-4a75-a9f9-d403684efeec\") " pod="metallb-system/metallb-operator-controller-manager-6887d9747b-t425t" Nov 29 04:49:37 crc kubenswrapper[4799]: I1129 04:49:37.061438 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pv2p7\" (UniqueName: \"kubernetes.io/projected/cabb2055-e31c-4a75-a9f9-d403684efeec-kube-api-access-pv2p7\") pod \"metallb-operator-controller-manager-6887d9747b-t425t\" (UID: \"cabb2055-e31c-4a75-a9f9-d403684efeec\") " pod="metallb-system/metallb-operator-controller-manager-6887d9747b-t425t" Nov 29 04:49:37 crc kubenswrapper[4799]: I1129 04:49:37.061485 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cabb2055-e31c-4a75-a9f9-d403684efeec-webhook-cert\") pod \"metallb-operator-controller-manager-6887d9747b-t425t\" (UID: \"cabb2055-e31c-4a75-a9f9-d403684efeec\") " pod="metallb-system/metallb-operator-controller-manager-6887d9747b-t425t" Nov 29 04:49:37 crc kubenswrapper[4799]: I1129 04:49:37.072706 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cabb2055-e31c-4a75-a9f9-d403684efeec-apiservice-cert\") pod \"metallb-operator-controller-manager-6887d9747b-t425t\" (UID: \"cabb2055-e31c-4a75-a9f9-d403684efeec\") " pod="metallb-system/metallb-operator-controller-manager-6887d9747b-t425t" Nov 29 04:49:37 crc kubenswrapper[4799]: I1129 04:49:37.082552 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cabb2055-e31c-4a75-a9f9-d403684efeec-webhook-cert\") pod \"metallb-operator-controller-manager-6887d9747b-t425t\" (UID: \"cabb2055-e31c-4a75-a9f9-d403684efeec\") " pod="metallb-system/metallb-operator-controller-manager-6887d9747b-t425t" Nov 29 04:49:37 crc kubenswrapper[4799]: I1129 04:49:37.108156 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pv2p7\" (UniqueName: \"kubernetes.io/projected/cabb2055-e31c-4a75-a9f9-d403684efeec-kube-api-access-pv2p7\") pod \"metallb-operator-controller-manager-6887d9747b-t425t\" (UID: \"cabb2055-e31c-4a75-a9f9-d403684efeec\") " pod="metallb-system/metallb-operator-controller-manager-6887d9747b-t425t" Nov 29 04:49:37 crc kubenswrapper[4799]: I1129 04:49:37.114224 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-78b9fb955c-gptcp"] Nov 29 04:49:37 crc kubenswrapper[4799]: I1129 04:49:37.115229 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-78b9fb955c-gptcp" Nov 29 04:49:37 crc kubenswrapper[4799]: I1129 04:49:37.117521 4799 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-fwsbb" Nov 29 04:49:37 crc kubenswrapper[4799]: I1129 04:49:37.118380 4799 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 29 04:49:37 crc kubenswrapper[4799]: I1129 04:49:37.118465 4799 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 29 04:49:37 crc kubenswrapper[4799]: I1129 04:49:37.137034 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-78b9fb955c-gptcp"] Nov 29 04:49:37 crc kubenswrapper[4799]: I1129 04:49:37.264194 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nll8s\" (UniqueName: \"kubernetes.io/projected/dd389a73-2aa5-4f69-8ff1-5a3e7a421c1e-kube-api-access-nll8s\") pod \"metallb-operator-webhook-server-78b9fb955c-gptcp\" (UID: \"dd389a73-2aa5-4f69-8ff1-5a3e7a421c1e\") " pod="metallb-system/metallb-operator-webhook-server-78b9fb955c-gptcp" Nov 29 04:49:37 crc kubenswrapper[4799]: I1129 04:49:37.264263 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/dd389a73-2aa5-4f69-8ff1-5a3e7a421c1e-apiservice-cert\") pod \"metallb-operator-webhook-server-78b9fb955c-gptcp\" (UID: \"dd389a73-2aa5-4f69-8ff1-5a3e7a421c1e\") " pod="metallb-system/metallb-operator-webhook-server-78b9fb955c-gptcp" Nov 29 04:49:37 crc kubenswrapper[4799]: I1129 04:49:37.264296 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/dd389a73-2aa5-4f69-8ff1-5a3e7a421c1e-webhook-cert\") pod \"metallb-operator-webhook-server-78b9fb955c-gptcp\" (UID: \"dd389a73-2aa5-4f69-8ff1-5a3e7a421c1e\") " pod="metallb-system/metallb-operator-webhook-server-78b9fb955c-gptcp" Nov 29 04:49:37 crc kubenswrapper[4799]: I1129 04:49:37.366282 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nll8s\" (UniqueName: \"kubernetes.io/projected/dd389a73-2aa5-4f69-8ff1-5a3e7a421c1e-kube-api-access-nll8s\") pod \"metallb-operator-webhook-server-78b9fb955c-gptcp\" (UID: \"dd389a73-2aa5-4f69-8ff1-5a3e7a421c1e\") " pod="metallb-system/metallb-operator-webhook-server-78b9fb955c-gptcp" Nov 29 04:49:37 crc kubenswrapper[4799]: I1129 04:49:37.366366 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/dd389a73-2aa5-4f69-8ff1-5a3e7a421c1e-apiservice-cert\") pod \"metallb-operator-webhook-server-78b9fb955c-gptcp\" (UID: \"dd389a73-2aa5-4f69-8ff1-5a3e7a421c1e\") " pod="metallb-system/metallb-operator-webhook-server-78b9fb955c-gptcp" Nov 29 04:49:37 crc kubenswrapper[4799]: I1129 04:49:37.366404 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/dd389a73-2aa5-4f69-8ff1-5a3e7a421c1e-webhook-cert\") pod \"metallb-operator-webhook-server-78b9fb955c-gptcp\" (UID: \"dd389a73-2aa5-4f69-8ff1-5a3e7a421c1e\") " pod="metallb-system/metallb-operator-webhook-server-78b9fb955c-gptcp" Nov 29 04:49:37 crc kubenswrapper[4799]: I1129 04:49:37.370048 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/dd389a73-2aa5-4f69-8ff1-5a3e7a421c1e-apiservice-cert\") pod \"metallb-operator-webhook-server-78b9fb955c-gptcp\" (UID: \"dd389a73-2aa5-4f69-8ff1-5a3e7a421c1e\") " pod="metallb-system/metallb-operator-webhook-server-78b9fb955c-gptcp" Nov 29 04:49:37 crc kubenswrapper[4799]: I1129 04:49:37.375747 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/dd389a73-2aa5-4f69-8ff1-5a3e7a421c1e-webhook-cert\") pod \"metallb-operator-webhook-server-78b9fb955c-gptcp\" (UID: \"dd389a73-2aa5-4f69-8ff1-5a3e7a421c1e\") " pod="metallb-system/metallb-operator-webhook-server-78b9fb955c-gptcp" Nov 29 04:49:37 crc kubenswrapper[4799]: I1129 04:49:37.388068 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nll8s\" (UniqueName: \"kubernetes.io/projected/dd389a73-2aa5-4f69-8ff1-5a3e7a421c1e-kube-api-access-nll8s\") pod \"metallb-operator-webhook-server-78b9fb955c-gptcp\" (UID: \"dd389a73-2aa5-4f69-8ff1-5a3e7a421c1e\") " pod="metallb-system/metallb-operator-webhook-server-78b9fb955c-gptcp" Nov 29 04:49:37 crc kubenswrapper[4799]: I1129 04:49:37.393560 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6887d9747b-t425t" Nov 29 04:49:37 crc kubenswrapper[4799]: I1129 04:49:37.458357 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-78b9fb955c-gptcp" Nov 29 04:49:37 crc kubenswrapper[4799]: I1129 04:49:37.804413 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-78b9fb955c-gptcp"] Nov 29 04:49:37 crc kubenswrapper[4799]: W1129 04:49:37.828526 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddd389a73_2aa5_4f69_8ff1_5a3e7a421c1e.slice/crio-7dc89ea7f9fca731f195218a21f7a0d185e59834296ca843e9275a4d0138ed79 WatchSource:0}: Error finding container 7dc89ea7f9fca731f195218a21f7a0d185e59834296ca843e9275a4d0138ed79: Status 404 returned error can't find the container with id 7dc89ea7f9fca731f195218a21f7a0d185e59834296ca843e9275a4d0138ed79 Nov 29 04:49:37 crc kubenswrapper[4799]: I1129 04:49:37.941058 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6887d9747b-t425t"] Nov 29 04:49:37 crc kubenswrapper[4799]: W1129 04:49:37.950002 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcabb2055_e31c_4a75_a9f9_d403684efeec.slice/crio-3ff82d47793060ae236f5beee9177f283475a4f331e308a2f257fa823faad16d WatchSource:0}: Error finding container 3ff82d47793060ae236f5beee9177f283475a4f331e308a2f257fa823faad16d: Status 404 returned error can't find the container with id 3ff82d47793060ae236f5beee9177f283475a4f331e308a2f257fa823faad16d Nov 29 04:49:38 crc kubenswrapper[4799]: I1129 04:49:38.683302 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6887d9747b-t425t" event={"ID":"cabb2055-e31c-4a75-a9f9-d403684efeec","Type":"ContainerStarted","Data":"3ff82d47793060ae236f5beee9177f283475a4f331e308a2f257fa823faad16d"} Nov 29 04:49:38 crc kubenswrapper[4799]: I1129 04:49:38.684484 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-78b9fb955c-gptcp" event={"ID":"dd389a73-2aa5-4f69-8ff1-5a3e7a421c1e","Type":"ContainerStarted","Data":"7dc89ea7f9fca731f195218a21f7a0d185e59834296ca843e9275a4d0138ed79"} Nov 29 04:49:43 crc kubenswrapper[4799]: I1129 04:49:43.720598 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6887d9747b-t425t" event={"ID":"cabb2055-e31c-4a75-a9f9-d403684efeec","Type":"ContainerStarted","Data":"d8eec20a23fbf6748fa02a38c318129f55742dc93555d335e07f3b2d2d60591e"} Nov 29 04:49:43 crc kubenswrapper[4799]: I1129 04:49:43.721564 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-6887d9747b-t425t" Nov 29 04:49:43 crc kubenswrapper[4799]: I1129 04:49:43.722616 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-78b9fb955c-gptcp" event={"ID":"dd389a73-2aa5-4f69-8ff1-5a3e7a421c1e","Type":"ContainerStarted","Data":"105114b62b337ce721f5965e7a5e7a51172d871d73c8e37ed0d84f83c52d6fc4"} Nov 29 04:49:43 crc kubenswrapper[4799]: I1129 04:49:43.722892 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-78b9fb955c-gptcp" Nov 29 04:49:43 crc kubenswrapper[4799]: I1129 04:49:43.751064 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-6887d9747b-t425t" podStartSLOduration=2.566240509 podStartE2EDuration="7.75103339s" podCreationTimestamp="2025-11-29 04:49:36 +0000 UTC" firstStartedPulling="2025-11-29 04:49:37.953332188 +0000 UTC m=+653.596262588" lastFinishedPulling="2025-11-29 04:49:43.138125069 +0000 UTC m=+658.781055469" observedRunningTime="2025-11-29 04:49:43.745645767 +0000 UTC m=+659.388576167" watchObservedRunningTime="2025-11-29 04:49:43.75103339 +0000 UTC m=+659.393963790" Nov 29 04:49:43 crc kubenswrapper[4799]: I1129 04:49:43.780194 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-78b9fb955c-gptcp" podStartSLOduration=1.440027545 podStartE2EDuration="6.780172474s" podCreationTimestamp="2025-11-29 04:49:37 +0000 UTC" firstStartedPulling="2025-11-29 04:49:37.831597226 +0000 UTC m=+653.474527616" lastFinishedPulling="2025-11-29 04:49:43.171742145 +0000 UTC m=+658.814672545" observedRunningTime="2025-11-29 04:49:43.774966365 +0000 UTC m=+659.417896765" watchObservedRunningTime="2025-11-29 04:49:43.780172474 +0000 UTC m=+659.423102874" Nov 29 04:49:57 crc kubenswrapper[4799]: I1129 04:49:57.464725 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-78b9fb955c-gptcp" Nov 29 04:50:17 crc kubenswrapper[4799]: I1129 04:50:17.397991 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-6887d9747b-t425t" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.165158 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-j9pvr"] Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.166237 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-j9pvr" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.173159 4799 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-jnfwl" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.175312 4799 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.178005 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-c95pk"] Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.181460 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-c95pk" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.184451 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.187037 4799 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.191076 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-j9pvr"] Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.242991 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xf26f\" (UniqueName: \"kubernetes.io/projected/bae7578a-afd9-462d-b824-beae428d9edf-kube-api-access-xf26f\") pod \"frr-k8s-c95pk\" (UID: \"bae7578a-afd9-462d-b824-beae428d9edf\") " pod="metallb-system/frr-k8s-c95pk" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.243065 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/bae7578a-afd9-462d-b824-beae428d9edf-frr-conf\") pod \"frr-k8s-c95pk\" (UID: \"bae7578a-afd9-462d-b824-beae428d9edf\") " pod="metallb-system/frr-k8s-c95pk" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.243118 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/95a0cc49-7fb2-4ed4-85a5-50d4248f07aa-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-j9pvr\" (UID: \"95a0cc49-7fb2-4ed4-85a5-50d4248f07aa\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-j9pvr" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.243155 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/bae7578a-afd9-462d-b824-beae428d9edf-metrics\") pod \"frr-k8s-c95pk\" (UID: \"bae7578a-afd9-462d-b824-beae428d9edf\") " pod="metallb-system/frr-k8s-c95pk" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.243184 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5xks\" (UniqueName: \"kubernetes.io/projected/95a0cc49-7fb2-4ed4-85a5-50d4248f07aa-kube-api-access-n5xks\") pod \"frr-k8s-webhook-server-7fcb986d4-j9pvr\" (UID: \"95a0cc49-7fb2-4ed4-85a5-50d4248f07aa\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-j9pvr" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.243257 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bae7578a-afd9-462d-b824-beae428d9edf-metrics-certs\") pod \"frr-k8s-c95pk\" (UID: \"bae7578a-afd9-462d-b824-beae428d9edf\") " pod="metallb-system/frr-k8s-c95pk" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.243305 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/bae7578a-afd9-462d-b824-beae428d9edf-reloader\") pod \"frr-k8s-c95pk\" (UID: \"bae7578a-afd9-462d-b824-beae428d9edf\") " pod="metallb-system/frr-k8s-c95pk" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.243341 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/bae7578a-afd9-462d-b824-beae428d9edf-frr-sockets\") pod \"frr-k8s-c95pk\" (UID: \"bae7578a-afd9-462d-b824-beae428d9edf\") " pod="metallb-system/frr-k8s-c95pk" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.243364 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/bae7578a-afd9-462d-b824-beae428d9edf-frr-startup\") pod \"frr-k8s-c95pk\" (UID: \"bae7578a-afd9-462d-b824-beae428d9edf\") " pod="metallb-system/frr-k8s-c95pk" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.269032 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-tc7fc"] Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.270072 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-tc7fc" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.275467 4799 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.275993 4799 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.276093 4799 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-6jwnh" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.276095 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.290780 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-27dxl"] Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.292117 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-27dxl" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.295095 4799 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.311631 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-27dxl"] Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.344536 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/0845b467-e744-408a-b2e7-d61b027724dd-metallb-excludel2\") pod \"speaker-tc7fc\" (UID: \"0845b467-e744-408a-b2e7-d61b027724dd\") " pod="metallb-system/speaker-tc7fc" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.344620 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/95a0cc49-7fb2-4ed4-85a5-50d4248f07aa-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-j9pvr\" (UID: \"95a0cc49-7fb2-4ed4-85a5-50d4248f07aa\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-j9pvr" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.344663 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4ztj\" (UniqueName: \"kubernetes.io/projected/9b8b8dfd-90f2-44b4-921a-e56563c76655-kube-api-access-k4ztj\") pod \"controller-f8648f98b-27dxl\" (UID: \"9b8b8dfd-90f2-44b4-921a-e56563c76655\") " pod="metallb-system/controller-f8648f98b-27dxl" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.344692 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/bae7578a-afd9-462d-b824-beae428d9edf-metrics\") pod \"frr-k8s-c95pk\" (UID: \"bae7578a-afd9-462d-b824-beae428d9edf\") " pod="metallb-system/frr-k8s-c95pk" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.344724 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5xks\" (UniqueName: \"kubernetes.io/projected/95a0cc49-7fb2-4ed4-85a5-50d4248f07aa-kube-api-access-n5xks\") pod \"frr-k8s-webhook-server-7fcb986d4-j9pvr\" (UID: \"95a0cc49-7fb2-4ed4-85a5-50d4248f07aa\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-j9pvr" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.344746 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9b8b8dfd-90f2-44b4-921a-e56563c76655-cert\") pod \"controller-f8648f98b-27dxl\" (UID: \"9b8b8dfd-90f2-44b4-921a-e56563c76655\") " pod="metallb-system/controller-f8648f98b-27dxl" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.344804 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hk6nb\" (UniqueName: \"kubernetes.io/projected/0845b467-e744-408a-b2e7-d61b027724dd-kube-api-access-hk6nb\") pod \"speaker-tc7fc\" (UID: \"0845b467-e744-408a-b2e7-d61b027724dd\") " pod="metallb-system/speaker-tc7fc" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.344830 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bae7578a-afd9-462d-b824-beae428d9edf-metrics-certs\") pod \"frr-k8s-c95pk\" (UID: \"bae7578a-afd9-462d-b824-beae428d9edf\") " pod="metallb-system/frr-k8s-c95pk" Nov 29 04:50:18 crc kubenswrapper[4799]: E1129 04:50:18.344844 4799 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Nov 29 04:50:18 crc kubenswrapper[4799]: E1129 04:50:18.344940 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/95a0cc49-7fb2-4ed4-85a5-50d4248f07aa-cert podName:95a0cc49-7fb2-4ed4-85a5-50d4248f07aa nodeName:}" failed. No retries permitted until 2025-11-29 04:50:18.844915193 +0000 UTC m=+694.487845593 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/95a0cc49-7fb2-4ed4-85a5-50d4248f07aa-cert") pod "frr-k8s-webhook-server-7fcb986d4-j9pvr" (UID: "95a0cc49-7fb2-4ed4-85a5-50d4248f07aa") : secret "frr-k8s-webhook-server-cert" not found Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.345181 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/bae7578a-afd9-462d-b824-beae428d9edf-metrics\") pod \"frr-k8s-c95pk\" (UID: \"bae7578a-afd9-462d-b824-beae428d9edf\") " pod="metallb-system/frr-k8s-c95pk" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.344858 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/bae7578a-afd9-462d-b824-beae428d9edf-reloader\") pod \"frr-k8s-c95pk\" (UID: \"bae7578a-afd9-462d-b824-beae428d9edf\") " pod="metallb-system/frr-k8s-c95pk" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.345295 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/bae7578a-afd9-462d-b824-beae428d9edf-reloader\") pod \"frr-k8s-c95pk\" (UID: \"bae7578a-afd9-462d-b824-beae428d9edf\") " pod="metallb-system/frr-k8s-c95pk" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.345298 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9b8b8dfd-90f2-44b4-921a-e56563c76655-metrics-certs\") pod \"controller-f8648f98b-27dxl\" (UID: \"9b8b8dfd-90f2-44b4-921a-e56563c76655\") " pod="metallb-system/controller-f8648f98b-27dxl" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.346876 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/bae7578a-afd9-462d-b824-beae428d9edf-frr-sockets\") pod \"frr-k8s-c95pk\" (UID: \"bae7578a-afd9-462d-b824-beae428d9edf\") " pod="metallb-system/frr-k8s-c95pk" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.346928 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/bae7578a-afd9-462d-b824-beae428d9edf-frr-startup\") pod \"frr-k8s-c95pk\" (UID: \"bae7578a-afd9-462d-b824-beae428d9edf\") " pod="metallb-system/frr-k8s-c95pk" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.347015 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0845b467-e744-408a-b2e7-d61b027724dd-metrics-certs\") pod \"speaker-tc7fc\" (UID: \"0845b467-e744-408a-b2e7-d61b027724dd\") " pod="metallb-system/speaker-tc7fc" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.347049 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xf26f\" (UniqueName: \"kubernetes.io/projected/bae7578a-afd9-462d-b824-beae428d9edf-kube-api-access-xf26f\") pod \"frr-k8s-c95pk\" (UID: \"bae7578a-afd9-462d-b824-beae428d9edf\") " pod="metallb-system/frr-k8s-c95pk" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.347087 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/0845b467-e744-408a-b2e7-d61b027724dd-memberlist\") pod \"speaker-tc7fc\" (UID: \"0845b467-e744-408a-b2e7-d61b027724dd\") " pod="metallb-system/speaker-tc7fc" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.347120 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/bae7578a-afd9-462d-b824-beae428d9edf-frr-conf\") pod \"frr-k8s-c95pk\" (UID: \"bae7578a-afd9-462d-b824-beae428d9edf\") " pod="metallb-system/frr-k8s-c95pk" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.347487 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/bae7578a-afd9-462d-b824-beae428d9edf-frr-conf\") pod \"frr-k8s-c95pk\" (UID: \"bae7578a-afd9-462d-b824-beae428d9edf\") " pod="metallb-system/frr-k8s-c95pk" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.347729 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/bae7578a-afd9-462d-b824-beae428d9edf-frr-sockets\") pod \"frr-k8s-c95pk\" (UID: \"bae7578a-afd9-462d-b824-beae428d9edf\") " pod="metallb-system/frr-k8s-c95pk" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.349457 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/bae7578a-afd9-462d-b824-beae428d9edf-frr-startup\") pod \"frr-k8s-c95pk\" (UID: \"bae7578a-afd9-462d-b824-beae428d9edf\") " pod="metallb-system/frr-k8s-c95pk" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.357917 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bae7578a-afd9-462d-b824-beae428d9edf-metrics-certs\") pod \"frr-k8s-c95pk\" (UID: \"bae7578a-afd9-462d-b824-beae428d9edf\") " pod="metallb-system/frr-k8s-c95pk" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.372665 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5xks\" (UniqueName: \"kubernetes.io/projected/95a0cc49-7fb2-4ed4-85a5-50d4248f07aa-kube-api-access-n5xks\") pod \"frr-k8s-webhook-server-7fcb986d4-j9pvr\" (UID: \"95a0cc49-7fb2-4ed4-85a5-50d4248f07aa\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-j9pvr" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.383471 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xf26f\" (UniqueName: \"kubernetes.io/projected/bae7578a-afd9-462d-b824-beae428d9edf-kube-api-access-xf26f\") pod \"frr-k8s-c95pk\" (UID: \"bae7578a-afd9-462d-b824-beae428d9edf\") " pod="metallb-system/frr-k8s-c95pk" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.448480 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0845b467-e744-408a-b2e7-d61b027724dd-metrics-certs\") pod \"speaker-tc7fc\" (UID: \"0845b467-e744-408a-b2e7-d61b027724dd\") " pod="metallb-system/speaker-tc7fc" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.448549 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/0845b467-e744-408a-b2e7-d61b027724dd-memberlist\") pod \"speaker-tc7fc\" (UID: \"0845b467-e744-408a-b2e7-d61b027724dd\") " pod="metallb-system/speaker-tc7fc" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.448598 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/0845b467-e744-408a-b2e7-d61b027724dd-metallb-excludel2\") pod \"speaker-tc7fc\" (UID: \"0845b467-e744-408a-b2e7-d61b027724dd\") " pod="metallb-system/speaker-tc7fc" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.448675 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4ztj\" (UniqueName: \"kubernetes.io/projected/9b8b8dfd-90f2-44b4-921a-e56563c76655-kube-api-access-k4ztj\") pod \"controller-f8648f98b-27dxl\" (UID: \"9b8b8dfd-90f2-44b4-921a-e56563c76655\") " pod="metallb-system/controller-f8648f98b-27dxl" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.448698 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9b8b8dfd-90f2-44b4-921a-e56563c76655-cert\") pod \"controller-f8648f98b-27dxl\" (UID: \"9b8b8dfd-90f2-44b4-921a-e56563c76655\") " pod="metallb-system/controller-f8648f98b-27dxl" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.448740 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hk6nb\" (UniqueName: \"kubernetes.io/projected/0845b467-e744-408a-b2e7-d61b027724dd-kube-api-access-hk6nb\") pod \"speaker-tc7fc\" (UID: \"0845b467-e744-408a-b2e7-d61b027724dd\") " pod="metallb-system/speaker-tc7fc" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.448765 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9b8b8dfd-90f2-44b4-921a-e56563c76655-metrics-certs\") pod \"controller-f8648f98b-27dxl\" (UID: \"9b8b8dfd-90f2-44b4-921a-e56563c76655\") " pod="metallb-system/controller-f8648f98b-27dxl" Nov 29 04:50:18 crc kubenswrapper[4799]: E1129 04:50:18.450032 4799 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 29 04:50:18 crc kubenswrapper[4799]: E1129 04:50:18.450136 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0845b467-e744-408a-b2e7-d61b027724dd-memberlist podName:0845b467-e744-408a-b2e7-d61b027724dd nodeName:}" failed. No retries permitted until 2025-11-29 04:50:18.950115879 +0000 UTC m=+694.593046279 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/0845b467-e744-408a-b2e7-d61b027724dd-memberlist") pod "speaker-tc7fc" (UID: "0845b467-e744-408a-b2e7-d61b027724dd") : secret "metallb-memberlist" not found Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.450231 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/0845b467-e744-408a-b2e7-d61b027724dd-metallb-excludel2\") pod \"speaker-tc7fc\" (UID: \"0845b467-e744-408a-b2e7-d61b027724dd\") " pod="metallb-system/speaker-tc7fc" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.452625 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9b8b8dfd-90f2-44b4-921a-e56563c76655-metrics-certs\") pod \"controller-f8648f98b-27dxl\" (UID: \"9b8b8dfd-90f2-44b4-921a-e56563c76655\") " pod="metallb-system/controller-f8648f98b-27dxl" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.453519 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0845b467-e744-408a-b2e7-d61b027724dd-metrics-certs\") pod \"speaker-tc7fc\" (UID: \"0845b467-e744-408a-b2e7-d61b027724dd\") " pod="metallb-system/speaker-tc7fc" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.453960 4799 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.464069 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9b8b8dfd-90f2-44b4-921a-e56563c76655-cert\") pod \"controller-f8648f98b-27dxl\" (UID: \"9b8b8dfd-90f2-44b4-921a-e56563c76655\") " pod="metallb-system/controller-f8648f98b-27dxl" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.468624 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4ztj\" (UniqueName: \"kubernetes.io/projected/9b8b8dfd-90f2-44b4-921a-e56563c76655-kube-api-access-k4ztj\") pod \"controller-f8648f98b-27dxl\" (UID: \"9b8b8dfd-90f2-44b4-921a-e56563c76655\") " pod="metallb-system/controller-f8648f98b-27dxl" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.469638 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hk6nb\" (UniqueName: \"kubernetes.io/projected/0845b467-e744-408a-b2e7-d61b027724dd-kube-api-access-hk6nb\") pod \"speaker-tc7fc\" (UID: \"0845b467-e744-408a-b2e7-d61b027724dd\") " pod="metallb-system/speaker-tc7fc" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.500103 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-c95pk" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.606423 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-27dxl" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.826723 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-27dxl"] Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.857584 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/95a0cc49-7fb2-4ed4-85a5-50d4248f07aa-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-j9pvr\" (UID: \"95a0cc49-7fb2-4ed4-85a5-50d4248f07aa\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-j9pvr" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.863277 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/95a0cc49-7fb2-4ed4-85a5-50d4248f07aa-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-j9pvr\" (UID: \"95a0cc49-7fb2-4ed4-85a5-50d4248f07aa\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-j9pvr" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.959097 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/0845b467-e744-408a-b2e7-d61b027724dd-memberlist\") pod \"speaker-tc7fc\" (UID: \"0845b467-e744-408a-b2e7-d61b027724dd\") " pod="metallb-system/speaker-tc7fc" Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.959367 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-c95pk" event={"ID":"bae7578a-afd9-462d-b824-beae428d9edf","Type":"ContainerStarted","Data":"b1e3231055653c7e0b356ca23f07e64595c38dcff7a50178e94ec1e6dc39079f"} Nov 29 04:50:18 crc kubenswrapper[4799]: E1129 04:50:18.959488 4799 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 29 04:50:18 crc kubenswrapper[4799]: E1129 04:50:18.959736 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0845b467-e744-408a-b2e7-d61b027724dd-memberlist podName:0845b467-e744-408a-b2e7-d61b027724dd nodeName:}" failed. No retries permitted until 2025-11-29 04:50:19.959552735 +0000 UTC m=+695.602483145 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/0845b467-e744-408a-b2e7-d61b027724dd-memberlist") pod "speaker-tc7fc" (UID: "0845b467-e744-408a-b2e7-d61b027724dd") : secret "metallb-memberlist" not found Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.960935 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-27dxl" event={"ID":"9b8b8dfd-90f2-44b4-921a-e56563c76655","Type":"ContainerStarted","Data":"f0b2fb19ea1f145b9db2062e9a2cf3452d963d189e414c6fe1f7a351cf0339c9"} Nov 29 04:50:18 crc kubenswrapper[4799]: I1129 04:50:18.960965 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-27dxl" event={"ID":"9b8b8dfd-90f2-44b4-921a-e56563c76655","Type":"ContainerStarted","Data":"3669a2e1ac0db9f007539674904d61692eecc52760248b79152c12179df5f8f2"} Nov 29 04:50:19 crc kubenswrapper[4799]: I1129 04:50:19.088734 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-j9pvr" Nov 29 04:50:19 crc kubenswrapper[4799]: I1129 04:50:19.529527 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-j9pvr"] Nov 29 04:50:19 crc kubenswrapper[4799]: I1129 04:50:19.975006 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-j9pvr" event={"ID":"95a0cc49-7fb2-4ed4-85a5-50d4248f07aa","Type":"ContainerStarted","Data":"97b58208a2f16d2738634323ffbcb926d6df3f53b9d520ae32b8a23a0346b033"} Nov 29 04:50:19 crc kubenswrapper[4799]: I1129 04:50:19.979570 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/0845b467-e744-408a-b2e7-d61b027724dd-memberlist\") pod \"speaker-tc7fc\" (UID: \"0845b467-e744-408a-b2e7-d61b027724dd\") " pod="metallb-system/speaker-tc7fc" Nov 29 04:50:19 crc kubenswrapper[4799]: I1129 04:50:19.980191 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-27dxl" event={"ID":"9b8b8dfd-90f2-44b4-921a-e56563c76655","Type":"ContainerStarted","Data":"56712bbeea6d34d792b2d11ae3baf86a566cc17944a98e8251732b0792e9d782"} Nov 29 04:50:19 crc kubenswrapper[4799]: I1129 04:50:19.980873 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-27dxl" Nov 29 04:50:19 crc kubenswrapper[4799]: I1129 04:50:19.993602 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/0845b467-e744-408a-b2e7-d61b027724dd-memberlist\") pod \"speaker-tc7fc\" (UID: \"0845b467-e744-408a-b2e7-d61b027724dd\") " pod="metallb-system/speaker-tc7fc" Nov 29 04:50:20 crc kubenswrapper[4799]: I1129 04:50:20.007944 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-27dxl" podStartSLOduration=2.007912115 podStartE2EDuration="2.007912115s" podCreationTimestamp="2025-11-29 04:50:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:50:20.005599463 +0000 UTC m=+695.648529863" watchObservedRunningTime="2025-11-29 04:50:20.007912115 +0000 UTC m=+695.650842505" Nov 29 04:50:20 crc kubenswrapper[4799]: I1129 04:50:20.086308 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-tc7fc" Nov 29 04:50:21 crc kubenswrapper[4799]: I1129 04:50:21.011481 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-tc7fc" event={"ID":"0845b467-e744-408a-b2e7-d61b027724dd","Type":"ContainerStarted","Data":"656bec3afdcb07a0f2b68f486c4c2588b47eba09f72ab9b26bdb4a2da99f352a"} Nov 29 04:50:21 crc kubenswrapper[4799]: I1129 04:50:21.011969 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-tc7fc" event={"ID":"0845b467-e744-408a-b2e7-d61b027724dd","Type":"ContainerStarted","Data":"b64db79b6c9b27c98051fd23ca42d469dd77b78c5e87bfcf2826d9f89b2f7c60"} Nov 29 04:50:21 crc kubenswrapper[4799]: I1129 04:50:21.012023 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-tc7fc" event={"ID":"0845b467-e744-408a-b2e7-d61b027724dd","Type":"ContainerStarted","Data":"641939ed8f2621fec5846f6ee63a788f1e3fadbeccacd3a4dda58b6626de2ffd"} Nov 29 04:50:21 crc kubenswrapper[4799]: I1129 04:50:21.012230 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-tc7fc" Nov 29 04:50:21 crc kubenswrapper[4799]: I1129 04:50:21.033204 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-tc7fc" podStartSLOduration=3.033179682 podStartE2EDuration="3.033179682s" podCreationTimestamp="2025-11-29 04:50:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:50:21.032395444 +0000 UTC m=+696.675325844" watchObservedRunningTime="2025-11-29 04:50:21.033179682 +0000 UTC m=+696.676110082" Nov 29 04:50:27 crc kubenswrapper[4799]: I1129 04:50:27.076011 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-j9pvr" event={"ID":"95a0cc49-7fb2-4ed4-85a5-50d4248f07aa","Type":"ContainerStarted","Data":"164609f2c60837ddb15c3dc1ff11a7db7fa3af82c1ec86875d1345db59d4d72a"} Nov 29 04:50:27 crc kubenswrapper[4799]: I1129 04:50:27.077138 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-j9pvr" Nov 29 04:50:27 crc kubenswrapper[4799]: I1129 04:50:27.078376 4799 generic.go:334] "Generic (PLEG): container finished" podID="bae7578a-afd9-462d-b824-beae428d9edf" containerID="d9b8d5811d3c1d345a48f6ff2d8a32903c3e23483f0fb36374f39467bf9643b3" exitCode=0 Nov 29 04:50:27 crc kubenswrapper[4799]: I1129 04:50:27.078438 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-c95pk" event={"ID":"bae7578a-afd9-462d-b824-beae428d9edf","Type":"ContainerDied","Data":"d9b8d5811d3c1d345a48f6ff2d8a32903c3e23483f0fb36374f39467bf9643b3"} Nov 29 04:50:27 crc kubenswrapper[4799]: I1129 04:50:27.104454 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-j9pvr" podStartSLOduration=2.72977861 podStartE2EDuration="9.104424475s" podCreationTimestamp="2025-11-29 04:50:18 +0000 UTC" firstStartedPulling="2025-11-29 04:50:19.549770719 +0000 UTC m=+695.192701119" lastFinishedPulling="2025-11-29 04:50:25.924416574 +0000 UTC m=+701.567346984" observedRunningTime="2025-11-29 04:50:27.102305987 +0000 UTC m=+702.745236387" watchObservedRunningTime="2025-11-29 04:50:27.104424475 +0000 UTC m=+702.747354875" Nov 29 04:50:27 crc kubenswrapper[4799]: I1129 04:50:27.248781 4799 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 29 04:50:28 crc kubenswrapper[4799]: I1129 04:50:28.085276 4799 generic.go:334] "Generic (PLEG): container finished" podID="bae7578a-afd9-462d-b824-beae428d9edf" containerID="6f8867585438d69b529025a3ad57afc8a1728eb312ee71fcb6a25f3fe263435e" exitCode=0 Nov 29 04:50:28 crc kubenswrapper[4799]: I1129 04:50:28.085331 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-c95pk" event={"ID":"bae7578a-afd9-462d-b824-beae428d9edf","Type":"ContainerDied","Data":"6f8867585438d69b529025a3ad57afc8a1728eb312ee71fcb6a25f3fe263435e"} Nov 29 04:50:28 crc kubenswrapper[4799]: I1129 04:50:28.611138 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-27dxl" Nov 29 04:50:29 crc kubenswrapper[4799]: I1129 04:50:29.095360 4799 generic.go:334] "Generic (PLEG): container finished" podID="bae7578a-afd9-462d-b824-beae428d9edf" containerID="fc5c3a0f84c38ddb3560764324bb2957d0f19414f18a22f4d4549b6b10e84874" exitCode=0 Nov 29 04:50:29 crc kubenswrapper[4799]: I1129 04:50:29.095438 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-c95pk" event={"ID":"bae7578a-afd9-462d-b824-beae428d9edf","Type":"ContainerDied","Data":"fc5c3a0f84c38ddb3560764324bb2957d0f19414f18a22f4d4549b6b10e84874"} Nov 29 04:50:30 crc kubenswrapper[4799]: I1129 04:50:30.091838 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-tc7fc" Nov 29 04:50:30 crc kubenswrapper[4799]: I1129 04:50:30.106411 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-c95pk" event={"ID":"bae7578a-afd9-462d-b824-beae428d9edf","Type":"ContainerStarted","Data":"40a481348564f2ccf931c4d0bd8e5d281440b7b02956aba5a480014dce40f826"} Nov 29 04:50:30 crc kubenswrapper[4799]: I1129 04:50:30.106460 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-c95pk" event={"ID":"bae7578a-afd9-462d-b824-beae428d9edf","Type":"ContainerStarted","Data":"1499d29508cf6b99953ce70c56dfaaff2fe787874e163a86c14e595d02b20392"} Nov 29 04:50:30 crc kubenswrapper[4799]: I1129 04:50:30.106469 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-c95pk" event={"ID":"bae7578a-afd9-462d-b824-beae428d9edf","Type":"ContainerStarted","Data":"08659a675ff61b692394fd520b1cdcd4696e565b0697aa42403f075de54c9e35"} Nov 29 04:50:30 crc kubenswrapper[4799]: I1129 04:50:30.106479 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-c95pk" event={"ID":"bae7578a-afd9-462d-b824-beae428d9edf","Type":"ContainerStarted","Data":"b9660f7ee6a65cebe2b9f0a75e1e62cfb391d6ba415ebddcdb0d72fcaeaddbb7"} Nov 29 04:50:30 crc kubenswrapper[4799]: I1129 04:50:30.106487 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-c95pk" event={"ID":"bae7578a-afd9-462d-b824-beae428d9edf","Type":"ContainerStarted","Data":"0960ff68cd643ee2b19cc3029b4f695306ea8c68495e2c2913d8ca47274ff448"} Nov 29 04:50:31 crc kubenswrapper[4799]: I1129 04:50:31.129036 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-c95pk" event={"ID":"bae7578a-afd9-462d-b824-beae428d9edf","Type":"ContainerStarted","Data":"f19ef563d151942c92a92cdc531e8f0c62b7e795efafa47ba6c78a5eb3a3a42e"} Nov 29 04:50:31 crc kubenswrapper[4799]: I1129 04:50:31.130457 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-c95pk" Nov 29 04:50:31 crc kubenswrapper[4799]: I1129 04:50:31.157371 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-c95pk" podStartSLOduration=5.813108759 podStartE2EDuration="13.157347612s" podCreationTimestamp="2025-11-29 04:50:18 +0000 UTC" firstStartedPulling="2025-11-29 04:50:18.594881717 +0000 UTC m=+694.237812117" lastFinishedPulling="2025-11-29 04:50:25.93912055 +0000 UTC m=+701.582050970" observedRunningTime="2025-11-29 04:50:31.15376044 +0000 UTC m=+706.796690850" watchObservedRunningTime="2025-11-29 04:50:31.157347612 +0000 UTC m=+706.800278022" Nov 29 04:50:32 crc kubenswrapper[4799]: I1129 04:50:32.895833 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-26dwm"] Nov 29 04:50:32 crc kubenswrapper[4799]: I1129 04:50:32.897057 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-26dwm" Nov 29 04:50:32 crc kubenswrapper[4799]: I1129 04:50:32.900931 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-lmsrb" Nov 29 04:50:32 crc kubenswrapper[4799]: I1129 04:50:32.901175 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 29 04:50:32 crc kubenswrapper[4799]: I1129 04:50:32.902758 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 29 04:50:32 crc kubenswrapper[4799]: I1129 04:50:32.953450 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-26dwm"] Nov 29 04:50:33 crc kubenswrapper[4799]: I1129 04:50:33.097909 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5gr7\" (UniqueName: \"kubernetes.io/projected/3fa00de8-d3d4-4447-b347-4c2abb034a18-kube-api-access-m5gr7\") pod \"openstack-operator-index-26dwm\" (UID: \"3fa00de8-d3d4-4447-b347-4c2abb034a18\") " pod="openstack-operators/openstack-operator-index-26dwm" Nov 29 04:50:33 crc kubenswrapper[4799]: I1129 04:50:33.199605 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5gr7\" (UniqueName: \"kubernetes.io/projected/3fa00de8-d3d4-4447-b347-4c2abb034a18-kube-api-access-m5gr7\") pod \"openstack-operator-index-26dwm\" (UID: \"3fa00de8-d3d4-4447-b347-4c2abb034a18\") " pod="openstack-operators/openstack-operator-index-26dwm" Nov 29 04:50:33 crc kubenswrapper[4799]: I1129 04:50:33.222756 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5gr7\" (UniqueName: \"kubernetes.io/projected/3fa00de8-d3d4-4447-b347-4c2abb034a18-kube-api-access-m5gr7\") pod \"openstack-operator-index-26dwm\" (UID: \"3fa00de8-d3d4-4447-b347-4c2abb034a18\") " pod="openstack-operators/openstack-operator-index-26dwm" Nov 29 04:50:33 crc kubenswrapper[4799]: I1129 04:50:33.306766 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-26dwm" Nov 29 04:50:33 crc kubenswrapper[4799]: I1129 04:50:33.501006 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-c95pk" Nov 29 04:50:33 crc kubenswrapper[4799]: I1129 04:50:33.545698 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-c95pk" Nov 29 04:50:33 crc kubenswrapper[4799]: I1129 04:50:33.754496 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-26dwm"] Nov 29 04:50:33 crc kubenswrapper[4799]: W1129 04:50:33.759546 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3fa00de8_d3d4_4447_b347_4c2abb034a18.slice/crio-54ac98f26ba77b7c0e977ac7727970f4bb6bdf7705480956e9b88eb5f94a8332 WatchSource:0}: Error finding container 54ac98f26ba77b7c0e977ac7727970f4bb6bdf7705480956e9b88eb5f94a8332: Status 404 returned error can't find the container with id 54ac98f26ba77b7c0e977ac7727970f4bb6bdf7705480956e9b88eb5f94a8332 Nov 29 04:50:34 crc kubenswrapper[4799]: I1129 04:50:34.150071 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-26dwm" event={"ID":"3fa00de8-d3d4-4447-b347-4c2abb034a18","Type":"ContainerStarted","Data":"54ac98f26ba77b7c0e977ac7727970f4bb6bdf7705480956e9b88eb5f94a8332"} Nov 29 04:50:36 crc kubenswrapper[4799]: I1129 04:50:36.071313 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-26dwm"] Nov 29 04:50:36 crc kubenswrapper[4799]: I1129 04:50:36.698090 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-hr4n9"] Nov 29 04:50:36 crc kubenswrapper[4799]: I1129 04:50:36.700703 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-hr4n9" Nov 29 04:50:36 crc kubenswrapper[4799]: I1129 04:50:36.706862 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-hr4n9"] Nov 29 04:50:36 crc kubenswrapper[4799]: I1129 04:50:36.853920 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sq99q\" (UniqueName: \"kubernetes.io/projected/8a6a8fd6-7ba3-4457-9d23-b18414836f30-kube-api-access-sq99q\") pod \"openstack-operator-index-hr4n9\" (UID: \"8a6a8fd6-7ba3-4457-9d23-b18414836f30\") " pod="openstack-operators/openstack-operator-index-hr4n9" Nov 29 04:50:36 crc kubenswrapper[4799]: I1129 04:50:36.956164 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sq99q\" (UniqueName: \"kubernetes.io/projected/8a6a8fd6-7ba3-4457-9d23-b18414836f30-kube-api-access-sq99q\") pod \"openstack-operator-index-hr4n9\" (UID: \"8a6a8fd6-7ba3-4457-9d23-b18414836f30\") " pod="openstack-operators/openstack-operator-index-hr4n9" Nov 29 04:50:36 crc kubenswrapper[4799]: I1129 04:50:36.991170 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sq99q\" (UniqueName: \"kubernetes.io/projected/8a6a8fd6-7ba3-4457-9d23-b18414836f30-kube-api-access-sq99q\") pod \"openstack-operator-index-hr4n9\" (UID: \"8a6a8fd6-7ba3-4457-9d23-b18414836f30\") " pod="openstack-operators/openstack-operator-index-hr4n9" Nov 29 04:50:37 crc kubenswrapper[4799]: I1129 04:50:37.048642 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-hr4n9" Nov 29 04:50:37 crc kubenswrapper[4799]: I1129 04:50:37.497650 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 04:50:37 crc kubenswrapper[4799]: I1129 04:50:37.497755 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 04:50:38 crc kubenswrapper[4799]: I1129 04:50:38.369852 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-hr4n9"] Nov 29 04:50:38 crc kubenswrapper[4799]: W1129 04:50:38.373978 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8a6a8fd6_7ba3_4457_9d23_b18414836f30.slice/crio-7dda88626d9070b0e343d7c4062bdb8481451e953def1083e2c52d5871e4b631 WatchSource:0}: Error finding container 7dda88626d9070b0e343d7c4062bdb8481451e953def1083e2c52d5871e4b631: Status 404 returned error can't find the container with id 7dda88626d9070b0e343d7c4062bdb8481451e953def1083e2c52d5871e4b631 Nov 29 04:50:39 crc kubenswrapper[4799]: I1129 04:50:39.093840 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-j9pvr" Nov 29 04:50:39 crc kubenswrapper[4799]: I1129 04:50:39.205183 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-26dwm" event={"ID":"3fa00de8-d3d4-4447-b347-4c2abb034a18","Type":"ContainerStarted","Data":"4ea57b24d10d47ad812bf38a29903a882dedfc6e6901903ece38bd7cee6bb7bb"} Nov 29 04:50:39 crc kubenswrapper[4799]: I1129 04:50:39.205358 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-26dwm" podUID="3fa00de8-d3d4-4447-b347-4c2abb034a18" containerName="registry-server" containerID="cri-o://4ea57b24d10d47ad812bf38a29903a882dedfc6e6901903ece38bd7cee6bb7bb" gracePeriod=2 Nov 29 04:50:39 crc kubenswrapper[4799]: I1129 04:50:39.207469 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-hr4n9" event={"ID":"8a6a8fd6-7ba3-4457-9d23-b18414836f30","Type":"ContainerStarted","Data":"f89977bef3d90caf2ad6684bafc933f3f03a92669a7e2d3ef11f83da075d0451"} Nov 29 04:50:39 crc kubenswrapper[4799]: I1129 04:50:39.207523 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-hr4n9" event={"ID":"8a6a8fd6-7ba3-4457-9d23-b18414836f30","Type":"ContainerStarted","Data":"7dda88626d9070b0e343d7c4062bdb8481451e953def1083e2c52d5871e4b631"} Nov 29 04:50:39 crc kubenswrapper[4799]: I1129 04:50:39.235473 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-26dwm" podStartSLOduration=2.806695242 podStartE2EDuration="7.235449099s" podCreationTimestamp="2025-11-29 04:50:32 +0000 UTC" firstStartedPulling="2025-11-29 04:50:33.763402988 +0000 UTC m=+709.406333388" lastFinishedPulling="2025-11-29 04:50:38.192156845 +0000 UTC m=+713.835087245" observedRunningTime="2025-11-29 04:50:39.234501846 +0000 UTC m=+714.877432236" watchObservedRunningTime="2025-11-29 04:50:39.235449099 +0000 UTC m=+714.878379499" Nov 29 04:50:39 crc kubenswrapper[4799]: I1129 04:50:39.259473 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-hr4n9" podStartSLOduration=3.203330661 podStartE2EDuration="3.259443556s" podCreationTimestamp="2025-11-29 04:50:36 +0000 UTC" firstStartedPulling="2025-11-29 04:50:38.378853256 +0000 UTC m=+714.021783656" lastFinishedPulling="2025-11-29 04:50:38.434966151 +0000 UTC m=+714.077896551" observedRunningTime="2025-11-29 04:50:39.2574407 +0000 UTC m=+714.900371110" watchObservedRunningTime="2025-11-29 04:50:39.259443556 +0000 UTC m=+714.902373966" Nov 29 04:50:39 crc kubenswrapper[4799]: I1129 04:50:39.541892 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-26dwm" Nov 29 04:50:39 crc kubenswrapper[4799]: I1129 04:50:39.698393 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m5gr7\" (UniqueName: \"kubernetes.io/projected/3fa00de8-d3d4-4447-b347-4c2abb034a18-kube-api-access-m5gr7\") pod \"3fa00de8-d3d4-4447-b347-4c2abb034a18\" (UID: \"3fa00de8-d3d4-4447-b347-4c2abb034a18\") " Nov 29 04:50:39 crc kubenswrapper[4799]: I1129 04:50:39.703951 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fa00de8-d3d4-4447-b347-4c2abb034a18-kube-api-access-m5gr7" (OuterVolumeSpecName: "kube-api-access-m5gr7") pod "3fa00de8-d3d4-4447-b347-4c2abb034a18" (UID: "3fa00de8-d3d4-4447-b347-4c2abb034a18"). InnerVolumeSpecName "kube-api-access-m5gr7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:50:39 crc kubenswrapper[4799]: I1129 04:50:39.800314 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m5gr7\" (UniqueName: \"kubernetes.io/projected/3fa00de8-d3d4-4447-b347-4c2abb034a18-kube-api-access-m5gr7\") on node \"crc\" DevicePath \"\"" Nov 29 04:50:40 crc kubenswrapper[4799]: I1129 04:50:40.216309 4799 generic.go:334] "Generic (PLEG): container finished" podID="3fa00de8-d3d4-4447-b347-4c2abb034a18" containerID="4ea57b24d10d47ad812bf38a29903a882dedfc6e6901903ece38bd7cee6bb7bb" exitCode=0 Nov 29 04:50:40 crc kubenswrapper[4799]: I1129 04:50:40.216387 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-26dwm" Nov 29 04:50:40 crc kubenswrapper[4799]: I1129 04:50:40.216423 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-26dwm" event={"ID":"3fa00de8-d3d4-4447-b347-4c2abb034a18","Type":"ContainerDied","Data":"4ea57b24d10d47ad812bf38a29903a882dedfc6e6901903ece38bd7cee6bb7bb"} Nov 29 04:50:40 crc kubenswrapper[4799]: I1129 04:50:40.216482 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-26dwm" event={"ID":"3fa00de8-d3d4-4447-b347-4c2abb034a18","Type":"ContainerDied","Data":"54ac98f26ba77b7c0e977ac7727970f4bb6bdf7705480956e9b88eb5f94a8332"} Nov 29 04:50:40 crc kubenswrapper[4799]: I1129 04:50:40.216507 4799 scope.go:117] "RemoveContainer" containerID="4ea57b24d10d47ad812bf38a29903a882dedfc6e6901903ece38bd7cee6bb7bb" Nov 29 04:50:40 crc kubenswrapper[4799]: I1129 04:50:40.237151 4799 scope.go:117] "RemoveContainer" containerID="4ea57b24d10d47ad812bf38a29903a882dedfc6e6901903ece38bd7cee6bb7bb" Nov 29 04:50:40 crc kubenswrapper[4799]: E1129 04:50:40.238930 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ea57b24d10d47ad812bf38a29903a882dedfc6e6901903ece38bd7cee6bb7bb\": container with ID starting with 4ea57b24d10d47ad812bf38a29903a882dedfc6e6901903ece38bd7cee6bb7bb not found: ID does not exist" containerID="4ea57b24d10d47ad812bf38a29903a882dedfc6e6901903ece38bd7cee6bb7bb" Nov 29 04:50:40 crc kubenswrapper[4799]: I1129 04:50:40.238995 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ea57b24d10d47ad812bf38a29903a882dedfc6e6901903ece38bd7cee6bb7bb"} err="failed to get container status \"4ea57b24d10d47ad812bf38a29903a882dedfc6e6901903ece38bd7cee6bb7bb\": rpc error: code = NotFound desc = could not find container \"4ea57b24d10d47ad812bf38a29903a882dedfc6e6901903ece38bd7cee6bb7bb\": container with ID starting with 4ea57b24d10d47ad812bf38a29903a882dedfc6e6901903ece38bd7cee6bb7bb not found: ID does not exist" Nov 29 04:50:40 crc kubenswrapper[4799]: I1129 04:50:40.254831 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-26dwm"] Nov 29 04:50:40 crc kubenswrapper[4799]: I1129 04:50:40.261871 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-26dwm"] Nov 29 04:50:40 crc kubenswrapper[4799]: I1129 04:50:40.667484 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3fa00de8-d3d4-4447-b347-4c2abb034a18" path="/var/lib/kubelet/pods/3fa00de8-d3d4-4447-b347-4c2abb034a18/volumes" Nov 29 04:50:47 crc kubenswrapper[4799]: I1129 04:50:47.049266 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-hr4n9" Nov 29 04:50:47 crc kubenswrapper[4799]: I1129 04:50:47.050337 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-hr4n9" Nov 29 04:50:47 crc kubenswrapper[4799]: I1129 04:50:47.092614 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-hr4n9" Nov 29 04:50:47 crc kubenswrapper[4799]: I1129 04:50:47.307438 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-hr4n9" Nov 29 04:50:48 crc kubenswrapper[4799]: I1129 04:50:48.504917 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-c95pk" Nov 29 04:50:48 crc kubenswrapper[4799]: I1129 04:50:48.533857 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/232b41aac13e001dc857b9795d2b854afae448fcc1728af84c1ef8a484ctr76"] Nov 29 04:50:48 crc kubenswrapper[4799]: E1129 04:50:48.534175 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fa00de8-d3d4-4447-b347-4c2abb034a18" containerName="registry-server" Nov 29 04:50:48 crc kubenswrapper[4799]: I1129 04:50:48.534197 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fa00de8-d3d4-4447-b347-4c2abb034a18" containerName="registry-server" Nov 29 04:50:48 crc kubenswrapper[4799]: I1129 04:50:48.534454 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fa00de8-d3d4-4447-b347-4c2abb034a18" containerName="registry-server" Nov 29 04:50:48 crc kubenswrapper[4799]: I1129 04:50:48.535582 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/232b41aac13e001dc857b9795d2b854afae448fcc1728af84c1ef8a484ctr76" Nov 29 04:50:48 crc kubenswrapper[4799]: I1129 04:50:48.541049 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-d8dlf" Nov 29 04:50:48 crc kubenswrapper[4799]: I1129 04:50:48.602416 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/232b41aac13e001dc857b9795d2b854afae448fcc1728af84c1ef8a484ctr76"] Nov 29 04:50:48 crc kubenswrapper[4799]: I1129 04:50:48.637194 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrzjf\" (UniqueName: \"kubernetes.io/projected/e3b7bcee-1d1a-41a2-b1ea-221fe74998ef-kube-api-access-nrzjf\") pod \"232b41aac13e001dc857b9795d2b854afae448fcc1728af84c1ef8a484ctr76\" (UID: \"e3b7bcee-1d1a-41a2-b1ea-221fe74998ef\") " pod="openstack-operators/232b41aac13e001dc857b9795d2b854afae448fcc1728af84c1ef8a484ctr76" Nov 29 04:50:48 crc kubenswrapper[4799]: I1129 04:50:48.637274 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e3b7bcee-1d1a-41a2-b1ea-221fe74998ef-bundle\") pod \"232b41aac13e001dc857b9795d2b854afae448fcc1728af84c1ef8a484ctr76\" (UID: \"e3b7bcee-1d1a-41a2-b1ea-221fe74998ef\") " pod="openstack-operators/232b41aac13e001dc857b9795d2b854afae448fcc1728af84c1ef8a484ctr76" Nov 29 04:50:48 crc kubenswrapper[4799]: I1129 04:50:48.637300 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e3b7bcee-1d1a-41a2-b1ea-221fe74998ef-util\") pod \"232b41aac13e001dc857b9795d2b854afae448fcc1728af84c1ef8a484ctr76\" (UID: \"e3b7bcee-1d1a-41a2-b1ea-221fe74998ef\") " pod="openstack-operators/232b41aac13e001dc857b9795d2b854afae448fcc1728af84c1ef8a484ctr76" Nov 29 04:50:48 crc kubenswrapper[4799]: I1129 04:50:48.739080 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrzjf\" (UniqueName: \"kubernetes.io/projected/e3b7bcee-1d1a-41a2-b1ea-221fe74998ef-kube-api-access-nrzjf\") pod \"232b41aac13e001dc857b9795d2b854afae448fcc1728af84c1ef8a484ctr76\" (UID: \"e3b7bcee-1d1a-41a2-b1ea-221fe74998ef\") " pod="openstack-operators/232b41aac13e001dc857b9795d2b854afae448fcc1728af84c1ef8a484ctr76" Nov 29 04:50:48 crc kubenswrapper[4799]: I1129 04:50:48.739162 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e3b7bcee-1d1a-41a2-b1ea-221fe74998ef-bundle\") pod \"232b41aac13e001dc857b9795d2b854afae448fcc1728af84c1ef8a484ctr76\" (UID: \"e3b7bcee-1d1a-41a2-b1ea-221fe74998ef\") " pod="openstack-operators/232b41aac13e001dc857b9795d2b854afae448fcc1728af84c1ef8a484ctr76" Nov 29 04:50:48 crc kubenswrapper[4799]: I1129 04:50:48.739192 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e3b7bcee-1d1a-41a2-b1ea-221fe74998ef-util\") pod \"232b41aac13e001dc857b9795d2b854afae448fcc1728af84c1ef8a484ctr76\" (UID: \"e3b7bcee-1d1a-41a2-b1ea-221fe74998ef\") " pod="openstack-operators/232b41aac13e001dc857b9795d2b854afae448fcc1728af84c1ef8a484ctr76" Nov 29 04:50:48 crc kubenswrapper[4799]: I1129 04:50:48.739946 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e3b7bcee-1d1a-41a2-b1ea-221fe74998ef-util\") pod \"232b41aac13e001dc857b9795d2b854afae448fcc1728af84c1ef8a484ctr76\" (UID: \"e3b7bcee-1d1a-41a2-b1ea-221fe74998ef\") " pod="openstack-operators/232b41aac13e001dc857b9795d2b854afae448fcc1728af84c1ef8a484ctr76" Nov 29 04:50:48 crc kubenswrapper[4799]: I1129 04:50:48.739967 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e3b7bcee-1d1a-41a2-b1ea-221fe74998ef-bundle\") pod \"232b41aac13e001dc857b9795d2b854afae448fcc1728af84c1ef8a484ctr76\" (UID: \"e3b7bcee-1d1a-41a2-b1ea-221fe74998ef\") " pod="openstack-operators/232b41aac13e001dc857b9795d2b854afae448fcc1728af84c1ef8a484ctr76" Nov 29 04:50:48 crc kubenswrapper[4799]: I1129 04:50:48.759698 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrzjf\" (UniqueName: \"kubernetes.io/projected/e3b7bcee-1d1a-41a2-b1ea-221fe74998ef-kube-api-access-nrzjf\") pod \"232b41aac13e001dc857b9795d2b854afae448fcc1728af84c1ef8a484ctr76\" (UID: \"e3b7bcee-1d1a-41a2-b1ea-221fe74998ef\") " pod="openstack-operators/232b41aac13e001dc857b9795d2b854afae448fcc1728af84c1ef8a484ctr76" Nov 29 04:50:48 crc kubenswrapper[4799]: I1129 04:50:48.863392 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/232b41aac13e001dc857b9795d2b854afae448fcc1728af84c1ef8a484ctr76" Nov 29 04:50:49 crc kubenswrapper[4799]: I1129 04:50:49.347832 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/232b41aac13e001dc857b9795d2b854afae448fcc1728af84c1ef8a484ctr76"] Nov 29 04:50:49 crc kubenswrapper[4799]: W1129 04:50:49.349279 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode3b7bcee_1d1a_41a2_b1ea_221fe74998ef.slice/crio-fef9884f90b20150f06ff3ea4865a0910363d72ddadce649fa81c16be82daa3b WatchSource:0}: Error finding container fef9884f90b20150f06ff3ea4865a0910363d72ddadce649fa81c16be82daa3b: Status 404 returned error can't find the container with id fef9884f90b20150f06ff3ea4865a0910363d72ddadce649fa81c16be82daa3b Nov 29 04:50:50 crc kubenswrapper[4799]: I1129 04:50:50.312450 4799 generic.go:334] "Generic (PLEG): container finished" podID="e3b7bcee-1d1a-41a2-b1ea-221fe74998ef" containerID="825445167ad73991c968caaf4d73a78568b25b9ef6f9b64693ec764603dc6c03" exitCode=0 Nov 29 04:50:50 crc kubenswrapper[4799]: I1129 04:50:50.312599 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/232b41aac13e001dc857b9795d2b854afae448fcc1728af84c1ef8a484ctr76" event={"ID":"e3b7bcee-1d1a-41a2-b1ea-221fe74998ef","Type":"ContainerDied","Data":"825445167ad73991c968caaf4d73a78568b25b9ef6f9b64693ec764603dc6c03"} Nov 29 04:50:50 crc kubenswrapper[4799]: I1129 04:50:50.312932 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/232b41aac13e001dc857b9795d2b854afae448fcc1728af84c1ef8a484ctr76" event={"ID":"e3b7bcee-1d1a-41a2-b1ea-221fe74998ef","Type":"ContainerStarted","Data":"fef9884f90b20150f06ff3ea4865a0910363d72ddadce649fa81c16be82daa3b"} Nov 29 04:50:51 crc kubenswrapper[4799]: I1129 04:50:51.322714 4799 generic.go:334] "Generic (PLEG): container finished" podID="e3b7bcee-1d1a-41a2-b1ea-221fe74998ef" containerID="d1df3a81bbad079c0b22c8313bb989079bd3c40b4f8171a034158c6dacd182e3" exitCode=0 Nov 29 04:50:51 crc kubenswrapper[4799]: I1129 04:50:51.322866 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/232b41aac13e001dc857b9795d2b854afae448fcc1728af84c1ef8a484ctr76" event={"ID":"e3b7bcee-1d1a-41a2-b1ea-221fe74998ef","Type":"ContainerDied","Data":"d1df3a81bbad079c0b22c8313bb989079bd3c40b4f8171a034158c6dacd182e3"} Nov 29 04:50:52 crc kubenswrapper[4799]: I1129 04:50:52.336673 4799 generic.go:334] "Generic (PLEG): container finished" podID="e3b7bcee-1d1a-41a2-b1ea-221fe74998ef" containerID="9bd99140f4e35c22446bc2de3c7f8293fb2c795f38fbcf57758933816abf89ec" exitCode=0 Nov 29 04:50:52 crc kubenswrapper[4799]: I1129 04:50:52.336846 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/232b41aac13e001dc857b9795d2b854afae448fcc1728af84c1ef8a484ctr76" event={"ID":"e3b7bcee-1d1a-41a2-b1ea-221fe74998ef","Type":"ContainerDied","Data":"9bd99140f4e35c22446bc2de3c7f8293fb2c795f38fbcf57758933816abf89ec"} Nov 29 04:50:53 crc kubenswrapper[4799]: I1129 04:50:53.686002 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/232b41aac13e001dc857b9795d2b854afae448fcc1728af84c1ef8a484ctr76" Nov 29 04:50:53 crc kubenswrapper[4799]: I1129 04:50:53.733539 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrzjf\" (UniqueName: \"kubernetes.io/projected/e3b7bcee-1d1a-41a2-b1ea-221fe74998ef-kube-api-access-nrzjf\") pod \"e3b7bcee-1d1a-41a2-b1ea-221fe74998ef\" (UID: \"e3b7bcee-1d1a-41a2-b1ea-221fe74998ef\") " Nov 29 04:50:53 crc kubenswrapper[4799]: I1129 04:50:53.733713 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e3b7bcee-1d1a-41a2-b1ea-221fe74998ef-bundle\") pod \"e3b7bcee-1d1a-41a2-b1ea-221fe74998ef\" (UID: \"e3b7bcee-1d1a-41a2-b1ea-221fe74998ef\") " Nov 29 04:50:53 crc kubenswrapper[4799]: I1129 04:50:53.733940 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e3b7bcee-1d1a-41a2-b1ea-221fe74998ef-util\") pod \"e3b7bcee-1d1a-41a2-b1ea-221fe74998ef\" (UID: \"e3b7bcee-1d1a-41a2-b1ea-221fe74998ef\") " Nov 29 04:50:53 crc kubenswrapper[4799]: I1129 04:50:53.737509 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3b7bcee-1d1a-41a2-b1ea-221fe74998ef-bundle" (OuterVolumeSpecName: "bundle") pod "e3b7bcee-1d1a-41a2-b1ea-221fe74998ef" (UID: "e3b7bcee-1d1a-41a2-b1ea-221fe74998ef"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:50:53 crc kubenswrapper[4799]: I1129 04:50:53.747556 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3b7bcee-1d1a-41a2-b1ea-221fe74998ef-kube-api-access-nrzjf" (OuterVolumeSpecName: "kube-api-access-nrzjf") pod "e3b7bcee-1d1a-41a2-b1ea-221fe74998ef" (UID: "e3b7bcee-1d1a-41a2-b1ea-221fe74998ef"). InnerVolumeSpecName "kube-api-access-nrzjf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:50:53 crc kubenswrapper[4799]: I1129 04:50:53.756513 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3b7bcee-1d1a-41a2-b1ea-221fe74998ef-util" (OuterVolumeSpecName: "util") pod "e3b7bcee-1d1a-41a2-b1ea-221fe74998ef" (UID: "e3b7bcee-1d1a-41a2-b1ea-221fe74998ef"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:50:53 crc kubenswrapper[4799]: I1129 04:50:53.836201 4799 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e3b7bcee-1d1a-41a2-b1ea-221fe74998ef-util\") on node \"crc\" DevicePath \"\"" Nov 29 04:50:53 crc kubenswrapper[4799]: I1129 04:50:53.836252 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrzjf\" (UniqueName: \"kubernetes.io/projected/e3b7bcee-1d1a-41a2-b1ea-221fe74998ef-kube-api-access-nrzjf\") on node \"crc\" DevicePath \"\"" Nov 29 04:50:53 crc kubenswrapper[4799]: I1129 04:50:53.836266 4799 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e3b7bcee-1d1a-41a2-b1ea-221fe74998ef-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 04:50:54 crc kubenswrapper[4799]: I1129 04:50:54.367647 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/232b41aac13e001dc857b9795d2b854afae448fcc1728af84c1ef8a484ctr76" event={"ID":"e3b7bcee-1d1a-41a2-b1ea-221fe74998ef","Type":"ContainerDied","Data":"fef9884f90b20150f06ff3ea4865a0910363d72ddadce649fa81c16be82daa3b"} Nov 29 04:50:54 crc kubenswrapper[4799]: I1129 04:50:54.368125 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fef9884f90b20150f06ff3ea4865a0910363d72ddadce649fa81c16be82daa3b" Nov 29 04:50:54 crc kubenswrapper[4799]: I1129 04:50:54.367727 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/232b41aac13e001dc857b9795d2b854afae448fcc1728af84c1ef8a484ctr76" Nov 29 04:51:00 crc kubenswrapper[4799]: I1129 04:51:00.781431 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-5bb85d56cf-bx84s"] Nov 29 04:51:00 crc kubenswrapper[4799]: E1129 04:51:00.782733 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3b7bcee-1d1a-41a2-b1ea-221fe74998ef" containerName="pull" Nov 29 04:51:00 crc kubenswrapper[4799]: I1129 04:51:00.782751 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3b7bcee-1d1a-41a2-b1ea-221fe74998ef" containerName="pull" Nov 29 04:51:00 crc kubenswrapper[4799]: E1129 04:51:00.782776 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3b7bcee-1d1a-41a2-b1ea-221fe74998ef" containerName="extract" Nov 29 04:51:00 crc kubenswrapper[4799]: I1129 04:51:00.782804 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3b7bcee-1d1a-41a2-b1ea-221fe74998ef" containerName="extract" Nov 29 04:51:00 crc kubenswrapper[4799]: E1129 04:51:00.782825 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3b7bcee-1d1a-41a2-b1ea-221fe74998ef" containerName="util" Nov 29 04:51:00 crc kubenswrapper[4799]: I1129 04:51:00.782834 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3b7bcee-1d1a-41a2-b1ea-221fe74998ef" containerName="util" Nov 29 04:51:00 crc kubenswrapper[4799]: I1129 04:51:00.783068 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3b7bcee-1d1a-41a2-b1ea-221fe74998ef" containerName="extract" Nov 29 04:51:00 crc kubenswrapper[4799]: I1129 04:51:00.783812 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-5bb85d56cf-bx84s" Nov 29 04:51:00 crc kubenswrapper[4799]: I1129 04:51:00.786708 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-lbnl9" Nov 29 04:51:00 crc kubenswrapper[4799]: I1129 04:51:00.815729 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-5bb85d56cf-bx84s"] Nov 29 04:51:00 crc kubenswrapper[4799]: I1129 04:51:00.855467 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxhvq\" (UniqueName: \"kubernetes.io/projected/70a5864d-82f9-416d-a713-8e1bfa00733d-kube-api-access-qxhvq\") pod \"openstack-operator-controller-operator-5bb85d56cf-bx84s\" (UID: \"70a5864d-82f9-416d-a713-8e1bfa00733d\") " pod="openstack-operators/openstack-operator-controller-operator-5bb85d56cf-bx84s" Nov 29 04:51:00 crc kubenswrapper[4799]: I1129 04:51:00.956938 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxhvq\" (UniqueName: \"kubernetes.io/projected/70a5864d-82f9-416d-a713-8e1bfa00733d-kube-api-access-qxhvq\") pod \"openstack-operator-controller-operator-5bb85d56cf-bx84s\" (UID: \"70a5864d-82f9-416d-a713-8e1bfa00733d\") " pod="openstack-operators/openstack-operator-controller-operator-5bb85d56cf-bx84s" Nov 29 04:51:00 crc kubenswrapper[4799]: I1129 04:51:00.981766 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxhvq\" (UniqueName: \"kubernetes.io/projected/70a5864d-82f9-416d-a713-8e1bfa00733d-kube-api-access-qxhvq\") pod \"openstack-operator-controller-operator-5bb85d56cf-bx84s\" (UID: \"70a5864d-82f9-416d-a713-8e1bfa00733d\") " pod="openstack-operators/openstack-operator-controller-operator-5bb85d56cf-bx84s" Nov 29 04:51:01 crc kubenswrapper[4799]: I1129 04:51:01.109566 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-5bb85d56cf-bx84s" Nov 29 04:51:01 crc kubenswrapper[4799]: I1129 04:51:01.362077 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-5bb85d56cf-bx84s"] Nov 29 04:51:01 crc kubenswrapper[4799]: I1129 04:51:01.427524 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-5bb85d56cf-bx84s" event={"ID":"70a5864d-82f9-416d-a713-8e1bfa00733d","Type":"ContainerStarted","Data":"cdcf5ff5e67f60fd338e83d2f97caa74adac242e6705a7a42abfcbfe684b67f9"} Nov 29 04:51:07 crc kubenswrapper[4799]: I1129 04:51:07.485083 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-5bb85d56cf-bx84s" event={"ID":"70a5864d-82f9-416d-a713-8e1bfa00733d","Type":"ContainerStarted","Data":"d48cf533597634b42baa689935fc457760e2b6070ba203316c276a44f32f6162"} Nov 29 04:51:07 crc kubenswrapper[4799]: I1129 04:51:07.485961 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-5bb85d56cf-bx84s" Nov 29 04:51:07 crc kubenswrapper[4799]: I1129 04:51:07.497839 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 04:51:07 crc kubenswrapper[4799]: I1129 04:51:07.497920 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 04:51:07 crc kubenswrapper[4799]: I1129 04:51:07.516244 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-5bb85d56cf-bx84s" podStartSLOduration=1.700790704 podStartE2EDuration="7.516212792s" podCreationTimestamp="2025-11-29 04:51:00 +0000 UTC" firstStartedPulling="2025-11-29 04:51:01.36954081 +0000 UTC m=+737.012471210" lastFinishedPulling="2025-11-29 04:51:07.184962898 +0000 UTC m=+742.827893298" observedRunningTime="2025-11-29 04:51:07.515463364 +0000 UTC m=+743.158393774" watchObservedRunningTime="2025-11-29 04:51:07.516212792 +0000 UTC m=+743.159143202" Nov 29 04:51:21 crc kubenswrapper[4799]: I1129 04:51:21.114076 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-5bb85d56cf-bx84s" Nov 29 04:51:37 crc kubenswrapper[4799]: I1129 04:51:37.497272 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 04:51:37 crc kubenswrapper[4799]: I1129 04:51:37.497898 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 04:51:37 crc kubenswrapper[4799]: I1129 04:51:37.497962 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-26t88" Nov 29 04:51:37 crc kubenswrapper[4799]: I1129 04:51:37.498672 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"99c5c08976d14d2fb8bfd60a883c0eaa07a2b4ec5ce6f9de4d8a5dd6e5c7212d"} pod="openshift-machine-config-operator/machine-config-daemon-26t88" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 29 04:51:37 crc kubenswrapper[4799]: I1129 04:51:37.498729 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" containerID="cri-o://99c5c08976d14d2fb8bfd60a883c0eaa07a2b4ec5ce6f9de4d8a5dd6e5c7212d" gracePeriod=600 Nov 29 04:51:38 crc kubenswrapper[4799]: I1129 04:51:38.692857 4799 generic.go:334] "Generic (PLEG): container finished" podID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerID="99c5c08976d14d2fb8bfd60a883c0eaa07a2b4ec5ce6f9de4d8a5dd6e5c7212d" exitCode=0 Nov 29 04:51:38 crc kubenswrapper[4799]: I1129 04:51:38.692938 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" event={"ID":"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8","Type":"ContainerDied","Data":"99c5c08976d14d2fb8bfd60a883c0eaa07a2b4ec5ce6f9de4d8a5dd6e5c7212d"} Nov 29 04:51:38 crc kubenswrapper[4799]: I1129 04:51:38.693949 4799 scope.go:117] "RemoveContainer" containerID="2b1b7386c0bb5e909eb4074926c911c29f96ed91593e4e76b3689a1397aeb0ff" Nov 29 04:51:39 crc kubenswrapper[4799]: I1129 04:51:39.704145 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" event={"ID":"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8","Type":"ContainerStarted","Data":"8a07429ac3c8e7bd45ed97e1ba2d6c50f73c78cd13fcb21a144bfb0b7bc995f6"} Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.047167 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-6vnhr"] Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.048275 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-6vnhr" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.051148 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-hstpd" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.057292 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-7zddj"] Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.059080 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-7zddj" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.061718 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-24xvf" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.067980 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-6vnhr"] Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.072954 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-7zddj"] Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.079116 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-pkrhh"] Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.081183 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-pkrhh" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.085330 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-nv42q" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.087211 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-pkrhh"] Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.126481 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-668d9c48b9-lkbfl"] Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.127642 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-lkbfl" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.130688 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-w5hjn" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.142697 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-668d9c48b9-lkbfl"] Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.149467 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skz66\" (UniqueName: \"kubernetes.io/projected/7232eadb-c887-440b-b6f9-bffa4a6dc871-kube-api-access-skz66\") pod \"barbican-operator-controller-manager-7d9dfd778-6vnhr\" (UID: \"7232eadb-c887-440b-b6f9-bffa4a6dc871\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-6vnhr" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.155467 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-p555g"] Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.156655 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-p555g" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.159635 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-xvsbl" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.179978 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kxgw8"] Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.181236 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kxgw8" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.185616 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-nct6c" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.193405 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kxgw8"] Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.201565 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-p555g"] Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.206844 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-pn7wx"] Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.207906 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-pn7wx" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.209977 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.210134 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-rs8wz" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.211501 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-25fzb"] Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.213961 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-25fzb" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.218299 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-mzldj" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.219727 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-pn7wx"] Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.224060 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-25fzb"] Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.231028 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-84f754f96-t5gjl"] Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.232142 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-84f754f96-t5gjl" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.236173 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-546d4bdf48-8p2z8"] Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.237175 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-8p2z8" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.241142 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-4nrqf" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.242811 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-84f754f96-t5gjl"] Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.246201 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-sfpmg"] Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.247221 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-sfpmg" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.247538 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-ssvh2" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.249770 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-5r66k" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.253664 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77s2s\" (UniqueName: \"kubernetes.io/projected/743ba96f-315f-4e0d-abb0-5a6389140c7b-kube-api-access-77s2s\") pod \"glance-operator-controller-manager-668d9c48b9-lkbfl\" (UID: \"743ba96f-315f-4e0d-abb0-5a6389140c7b\") " pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-lkbfl" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.253735 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jq9b\" (UniqueName: \"kubernetes.io/projected/89038368-29a3-417e-8ffa-819e5901a52a-kube-api-access-9jq9b\") pod \"designate-operator-controller-manager-78b4bc895b-pkrhh\" (UID: \"89038368-29a3-417e-8ffa-819e5901a52a\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-pkrhh" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.253821 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skz66\" (UniqueName: \"kubernetes.io/projected/7232eadb-c887-440b-b6f9-bffa4a6dc871-kube-api-access-skz66\") pod \"barbican-operator-controller-manager-7d9dfd778-6vnhr\" (UID: \"7232eadb-c887-440b-b6f9-bffa4a6dc871\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-6vnhr" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.253861 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6nng\" (UniqueName: \"kubernetes.io/projected/2798aeab-f9b9-4728-ad63-186b16afcc31-kube-api-access-j6nng\") pod \"heat-operator-controller-manager-5f64f6f8bb-p555g\" (UID: \"2798aeab-f9b9-4728-ad63-186b16afcc31\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-p555g" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.254130 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6hfv\" (UniqueName: \"kubernetes.io/projected/259abc1e-a30c-4a8b-acd1-18052862198e-kube-api-access-l6hfv\") pod \"cinder-operator-controller-manager-859b6ccc6-7zddj\" (UID: \"259abc1e-a30c-4a8b-acd1-18052862198e\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-7zddj" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.283413 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-546d4bdf48-8p2z8"] Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.293333 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skz66\" (UniqueName: \"kubernetes.io/projected/7232eadb-c887-440b-b6f9-bffa4a6dc871-kube-api-access-skz66\") pod \"barbican-operator-controller-manager-7d9dfd778-6vnhr\" (UID: \"7232eadb-c887-440b-b6f9-bffa4a6dc871\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-6vnhr" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.312013 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-pc2zl"] Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.313287 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-pc2zl" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.323483 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-m979f" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.338847 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-sfpmg"] Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.355209 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4xpw\" (UniqueName: \"kubernetes.io/projected/f696db58-bd41-4db0-9ec7-c4a4710e6b8e-kube-api-access-j4xpw\") pod \"keystone-operator-controller-manager-546d4bdf48-8p2z8\" (UID: \"f696db58-bd41-4db0-9ec7-c4a4710e6b8e\") " pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-8p2z8" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.355247 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjlr6\" (UniqueName: \"kubernetes.io/projected/4bf2a041-0f89-480e-88d2-ba630db189c1-kube-api-access-xjlr6\") pod \"manila-operator-controller-manager-84f754f96-t5gjl\" (UID: \"4bf2a041-0f89-480e-88d2-ba630db189c1\") " pod="openstack-operators/manila-operator-controller-manager-84f754f96-t5gjl" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.355302 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6nng\" (UniqueName: \"kubernetes.io/projected/2798aeab-f9b9-4728-ad63-186b16afcc31-kube-api-access-j6nng\") pod \"heat-operator-controller-manager-5f64f6f8bb-p555g\" (UID: \"2798aeab-f9b9-4728-ad63-186b16afcc31\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-p555g" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.355336 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8pz2\" (UniqueName: \"kubernetes.io/projected/94ec7bc8-8413-4040-90dd-a5c9ce4ca98f-kube-api-access-g8pz2\") pod \"horizon-operator-controller-manager-68c6d99b8f-kxgw8\" (UID: \"94ec7bc8-8413-4040-90dd-a5c9ce4ca98f\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kxgw8" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.355362 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8c9ec062-c2b9-47a6-8d7a-eb9f00e63ef3-cert\") pod \"infra-operator-controller-manager-57548d458d-pn7wx\" (UID: \"8c9ec062-c2b9-47a6-8d7a-eb9f00e63ef3\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-pn7wx" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.355387 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjn2w\" (UniqueName: \"kubernetes.io/projected/8c9ec062-c2b9-47a6-8d7a-eb9f00e63ef3-kube-api-access-pjn2w\") pod \"infra-operator-controller-manager-57548d458d-pn7wx\" (UID: \"8c9ec062-c2b9-47a6-8d7a-eb9f00e63ef3\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-pn7wx" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.355415 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6hfv\" (UniqueName: \"kubernetes.io/projected/259abc1e-a30c-4a8b-acd1-18052862198e-kube-api-access-l6hfv\") pod \"cinder-operator-controller-manager-859b6ccc6-7zddj\" (UID: \"259abc1e-a30c-4a8b-acd1-18052862198e\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-7zddj" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.355443 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96w2q\" (UniqueName: \"kubernetes.io/projected/7851df67-88d5-45d8-804e-5d3bf3fdb3af-kube-api-access-96w2q\") pod \"mariadb-operator-controller-manager-56bbcc9d85-sfpmg\" (UID: \"7851df67-88d5-45d8-804e-5d3bf3fdb3af\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-sfpmg" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.355482 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77s2s\" (UniqueName: \"kubernetes.io/projected/743ba96f-315f-4e0d-abb0-5a6389140c7b-kube-api-access-77s2s\") pod \"glance-operator-controller-manager-668d9c48b9-lkbfl\" (UID: \"743ba96f-315f-4e0d-abb0-5a6389140c7b\") " pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-lkbfl" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.355524 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5b5k\" (UniqueName: \"kubernetes.io/projected/ff72829c-dfd8-4ae2-b508-f89e40f654eb-kube-api-access-p5b5k\") pod \"ironic-operator-controller-manager-6c548fd776-25fzb\" (UID: \"ff72829c-dfd8-4ae2-b508-f89e40f654eb\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-25fzb" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.355569 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jq9b\" (UniqueName: \"kubernetes.io/projected/89038368-29a3-417e-8ffa-819e5901a52a-kube-api-access-9jq9b\") pod \"designate-operator-controller-manager-78b4bc895b-pkrhh\" (UID: \"89038368-29a3-417e-8ffa-819e5901a52a\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-pkrhh" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.359030 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-8nl87"] Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.361082 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-8nl87" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.371631 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-6vnhr" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.373304 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-vz95f" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.397930 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-pc2zl"] Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.411157 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77s2s\" (UniqueName: \"kubernetes.io/projected/743ba96f-315f-4e0d-abb0-5a6389140c7b-kube-api-access-77s2s\") pod \"glance-operator-controller-manager-668d9c48b9-lkbfl\" (UID: \"743ba96f-315f-4e0d-abb0-5a6389140c7b\") " pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-lkbfl" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.427733 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6nng\" (UniqueName: \"kubernetes.io/projected/2798aeab-f9b9-4728-ad63-186b16afcc31-kube-api-access-j6nng\") pod \"heat-operator-controller-manager-5f64f6f8bb-p555g\" (UID: \"2798aeab-f9b9-4728-ad63-186b16afcc31\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-p555g" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.436132 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-wstbs"] Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.436805 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jq9b\" (UniqueName: \"kubernetes.io/projected/89038368-29a3-417e-8ffa-819e5901a52a-kube-api-access-9jq9b\") pod \"designate-operator-controller-manager-78b4bc895b-pkrhh\" (UID: \"89038368-29a3-417e-8ffa-819e5901a52a\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-pkrhh" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.437366 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6hfv\" (UniqueName: \"kubernetes.io/projected/259abc1e-a30c-4a8b-acd1-18052862198e-kube-api-access-l6hfv\") pod \"cinder-operator-controller-manager-859b6ccc6-7zddj\" (UID: \"259abc1e-a30c-4a8b-acd1-18052862198e\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-7zddj" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.444481 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-wstbs" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.456297 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-mg8rb" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.457174 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-lkbfl" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.458710 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5b5k\" (UniqueName: \"kubernetes.io/projected/ff72829c-dfd8-4ae2-b508-f89e40f654eb-kube-api-access-p5b5k\") pod \"ironic-operator-controller-manager-6c548fd776-25fzb\" (UID: \"ff72829c-dfd8-4ae2-b508-f89e40f654eb\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-25fzb" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.461394 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4xpw\" (UniqueName: \"kubernetes.io/projected/f696db58-bd41-4db0-9ec7-c4a4710e6b8e-kube-api-access-j4xpw\") pod \"keystone-operator-controller-manager-546d4bdf48-8p2z8\" (UID: \"f696db58-bd41-4db0-9ec7-c4a4710e6b8e\") " pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-8p2z8" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.461592 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjlr6\" (UniqueName: \"kubernetes.io/projected/4bf2a041-0f89-480e-88d2-ba630db189c1-kube-api-access-xjlr6\") pod \"manila-operator-controller-manager-84f754f96-t5gjl\" (UID: \"4bf2a041-0f89-480e-88d2-ba630db189c1\") " pod="openstack-operators/manila-operator-controller-manager-84f754f96-t5gjl" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.463940 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8pz2\" (UniqueName: \"kubernetes.io/projected/94ec7bc8-8413-4040-90dd-a5c9ce4ca98f-kube-api-access-g8pz2\") pod \"horizon-operator-controller-manager-68c6d99b8f-kxgw8\" (UID: \"94ec7bc8-8413-4040-90dd-a5c9ce4ca98f\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kxgw8" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.464070 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8c9ec062-c2b9-47a6-8d7a-eb9f00e63ef3-cert\") pod \"infra-operator-controller-manager-57548d458d-pn7wx\" (UID: \"8c9ec062-c2b9-47a6-8d7a-eb9f00e63ef3\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-pn7wx" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.464186 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjn2w\" (UniqueName: \"kubernetes.io/projected/8c9ec062-c2b9-47a6-8d7a-eb9f00e63ef3-kube-api-access-pjn2w\") pod \"infra-operator-controller-manager-57548d458d-pn7wx\" (UID: \"8c9ec062-c2b9-47a6-8d7a-eb9f00e63ef3\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-pn7wx" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.465037 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96w2q\" (UniqueName: \"kubernetes.io/projected/7851df67-88d5-45d8-804e-5d3bf3fdb3af-kube-api-access-96w2q\") pod \"mariadb-operator-controller-manager-56bbcc9d85-sfpmg\" (UID: \"7851df67-88d5-45d8-804e-5d3bf3fdb3af\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-sfpmg" Nov 29 04:51:40 crc kubenswrapper[4799]: E1129 04:51:40.465981 4799 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 29 04:51:40 crc kubenswrapper[4799]: E1129 04:51:40.468182 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8c9ec062-c2b9-47a6-8d7a-eb9f00e63ef3-cert podName:8c9ec062-c2b9-47a6-8d7a-eb9f00e63ef3 nodeName:}" failed. No retries permitted until 2025-11-29 04:51:40.968143251 +0000 UTC m=+776.611073651 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/8c9ec062-c2b9-47a6-8d7a-eb9f00e63ef3-cert") pod "infra-operator-controller-manager-57548d458d-pn7wx" (UID: "8c9ec062-c2b9-47a6-8d7a-eb9f00e63ef3") : secret "infra-operator-webhook-server-cert" not found Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.469114 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bt8cm\" (UniqueName: \"kubernetes.io/projected/e1d949de-4ce7-4016-83fd-4c0ce003eb0f-kube-api-access-bt8cm\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-pc2zl\" (UID: \"e1d949de-4ce7-4016-83fd-4c0ce003eb0f\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-pc2zl" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.469334 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9qvz\" (UniqueName: \"kubernetes.io/projected/bdcce387-cd82-490e-bd50-cf47158b1ea6-kube-api-access-g9qvz\") pod \"nova-operator-controller-manager-697bc559fc-8nl87\" (UID: \"bdcce387-cd82-490e-bd50-cf47158b1ea6\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-8nl87" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.476210 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-p555g" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.485992 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-8nl87"] Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.488493 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8pz2\" (UniqueName: \"kubernetes.io/projected/94ec7bc8-8413-4040-90dd-a5c9ce4ca98f-kube-api-access-g8pz2\") pod \"horizon-operator-controller-manager-68c6d99b8f-kxgw8\" (UID: \"94ec7bc8-8413-4040-90dd-a5c9ce4ca98f\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kxgw8" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.488912 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4xpw\" (UniqueName: \"kubernetes.io/projected/f696db58-bd41-4db0-9ec7-c4a4710e6b8e-kube-api-access-j4xpw\") pod \"keystone-operator-controller-manager-546d4bdf48-8p2z8\" (UID: \"f696db58-bd41-4db0-9ec7-c4a4710e6b8e\") " pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-8p2z8" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.490176 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5b5k\" (UniqueName: \"kubernetes.io/projected/ff72829c-dfd8-4ae2-b508-f89e40f654eb-kube-api-access-p5b5k\") pod \"ironic-operator-controller-manager-6c548fd776-25fzb\" (UID: \"ff72829c-dfd8-4ae2-b508-f89e40f654eb\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-25fzb" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.497554 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjlr6\" (UniqueName: \"kubernetes.io/projected/4bf2a041-0f89-480e-88d2-ba630db189c1-kube-api-access-xjlr6\") pod \"manila-operator-controller-manager-84f754f96-t5gjl\" (UID: \"4bf2a041-0f89-480e-88d2-ba630db189c1\") " pod="openstack-operators/manila-operator-controller-manager-84f754f96-t5gjl" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.497782 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjn2w\" (UniqueName: \"kubernetes.io/projected/8c9ec062-c2b9-47a6-8d7a-eb9f00e63ef3-kube-api-access-pjn2w\") pod \"infra-operator-controller-manager-57548d458d-pn7wx\" (UID: \"8c9ec062-c2b9-47a6-8d7a-eb9f00e63ef3\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-pn7wx" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.498563 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96w2q\" (UniqueName: \"kubernetes.io/projected/7851df67-88d5-45d8-804e-5d3bf3fdb3af-kube-api-access-96w2q\") pod \"mariadb-operator-controller-manager-56bbcc9d85-sfpmg\" (UID: \"7851df67-88d5-45d8-804e-5d3bf3fdb3af\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-sfpmg" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.507984 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kxgw8" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.511578 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-wstbs"] Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.544898 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-25fzb" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.546871 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4qnwnt"] Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.550963 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4qnwnt" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.554209 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-lpq6g" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.554520 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.565125 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-84f754f96-t5gjl" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.572224 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnlg6\" (UniqueName: \"kubernetes.io/projected/bbe2e2fb-4619-405c-a9ad-4bdf727882cd-kube-api-access-jnlg6\") pod \"octavia-operator-controller-manager-998648c74-wstbs\" (UID: \"bbe2e2fb-4619-405c-a9ad-4bdf727882cd\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-wstbs" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.572352 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bt8cm\" (UniqueName: \"kubernetes.io/projected/e1d949de-4ce7-4016-83fd-4c0ce003eb0f-kube-api-access-bt8cm\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-pc2zl\" (UID: \"e1d949de-4ce7-4016-83fd-4c0ce003eb0f\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-pc2zl" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.572391 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9qvz\" (UniqueName: \"kubernetes.io/projected/bdcce387-cd82-490e-bd50-cf47158b1ea6-kube-api-access-g9qvz\") pod \"nova-operator-controller-manager-697bc559fc-8nl87\" (UID: \"bdcce387-cd82-490e-bd50-cf47158b1ea6\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-8nl87" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.574611 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-qfp99"] Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.577503 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-qfp99" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.592150 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-8p2z8" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.593469 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-qfp99"] Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.594657 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-wpsn7" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.595963 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-sfpmg" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.597670 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bt8cm\" (UniqueName: \"kubernetes.io/projected/e1d949de-4ce7-4016-83fd-4c0ce003eb0f-kube-api-access-bt8cm\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-pc2zl\" (UID: \"e1d949de-4ce7-4016-83fd-4c0ce003eb0f\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-pc2zl" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.598608 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4qnwnt"] Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.611516 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-4p69t"] Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.615416 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-4p69t" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.620018 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-s2qj9" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.620105 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9qvz\" (UniqueName: \"kubernetes.io/projected/bdcce387-cd82-490e-bd50-cf47158b1ea6-kube-api-access-g9qvz\") pod \"nova-operator-controller-manager-697bc559fc-8nl87\" (UID: \"bdcce387-cd82-490e-bd50-cf47158b1ea6\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-8nl87" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.625552 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-4p69t"] Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.652870 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-pc2zl" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.654146 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-4xvn4"] Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.655465 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-4xvn4" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.658679 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-p56qj" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.676538 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8e0677fe-7f53-4e50-881b-943fffe1c0ff-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4qnwnt\" (UID: \"8e0677fe-7f53-4e50-881b-943fffe1c0ff\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4qnwnt" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.676617 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w54bx\" (UniqueName: \"kubernetes.io/projected/8e0677fe-7f53-4e50-881b-943fffe1c0ff-kube-api-access-w54bx\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4qnwnt\" (UID: \"8e0677fe-7f53-4e50-881b-943fffe1c0ff\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4qnwnt" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.676689 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnlg6\" (UniqueName: \"kubernetes.io/projected/bbe2e2fb-4619-405c-a9ad-4bdf727882cd-kube-api-access-jnlg6\") pod \"octavia-operator-controller-manager-998648c74-wstbs\" (UID: \"bbe2e2fb-4619-405c-a9ad-4bdf727882cd\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-wstbs" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.680556 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-7zddj" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.687395 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-4xvn4"] Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.701535 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-pkrhh" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.711075 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-wv65n"] Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.711301 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnlg6\" (UniqueName: \"kubernetes.io/projected/bbe2e2fb-4619-405c-a9ad-4bdf727882cd-kube-api-access-jnlg6\") pod \"octavia-operator-controller-manager-998648c74-wstbs\" (UID: \"bbe2e2fb-4619-405c-a9ad-4bdf727882cd\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-wstbs" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.716195 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-wv65n" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.719976 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-zm2s2" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.724484 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-wv65n"] Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.738095 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-tbth4"] Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.739275 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-tbth4" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.743239 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-tbth4"] Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.752420 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-sm9d2" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.781537 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bvbq\" (UniqueName: \"kubernetes.io/projected/861aea23-001f-4259-b69a-64b95db49c73-kube-api-access-6bvbq\") pod \"placement-operator-controller-manager-78f8948974-4p69t\" (UID: \"861aea23-001f-4259-b69a-64b95db49c73\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-4p69t" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.781616 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8e0677fe-7f53-4e50-881b-943fffe1c0ff-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4qnwnt\" (UID: \"8e0677fe-7f53-4e50-881b-943fffe1c0ff\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4qnwnt" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.781639 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w54bx\" (UniqueName: \"kubernetes.io/projected/8e0677fe-7f53-4e50-881b-943fffe1c0ff-kube-api-access-w54bx\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4qnwnt\" (UID: \"8e0677fe-7f53-4e50-881b-943fffe1c0ff\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4qnwnt" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.781660 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zj6h\" (UniqueName: \"kubernetes.io/projected/888d5a2b-8a13-4042-8c43-59241cfdc088-kube-api-access-8zj6h\") pod \"ovn-operator-controller-manager-b6456fdb6-qfp99\" (UID: \"888d5a2b-8a13-4042-8c43-59241cfdc088\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-qfp99" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.781693 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdjm6\" (UniqueName: \"kubernetes.io/projected/ac90f9ba-d2ef-44a1-84e7-8e809b5eef13-kube-api-access-wdjm6\") pod \"swift-operator-controller-manager-5f8c65bbfc-4xvn4\" (UID: \"ac90f9ba-d2ef-44a1-84e7-8e809b5eef13\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-4xvn4" Nov 29 04:51:40 crc kubenswrapper[4799]: E1129 04:51:40.782287 4799 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 29 04:51:40 crc kubenswrapper[4799]: E1129 04:51:40.782348 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8e0677fe-7f53-4e50-881b-943fffe1c0ff-cert podName:8e0677fe-7f53-4e50-881b-943fffe1c0ff nodeName:}" failed. No retries permitted until 2025-11-29 04:51:41.282331079 +0000 UTC m=+776.925261479 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/8e0677fe-7f53-4e50-881b-943fffe1c0ff-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4qnwnt" (UID: "8e0677fe-7f53-4e50-881b-943fffe1c0ff") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.794833 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-8nl87" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.806099 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w54bx\" (UniqueName: \"kubernetes.io/projected/8e0677fe-7f53-4e50-881b-943fffe1c0ff-kube-api-access-w54bx\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4qnwnt\" (UID: \"8e0677fe-7f53-4e50-881b-943fffe1c0ff\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4qnwnt" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.806178 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-v4kv7"] Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.812883 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-v4kv7" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.823018 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-7f9v4" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.823885 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-v4kv7"] Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.840364 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-wstbs" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.884404 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmmnp\" (UniqueName: \"kubernetes.io/projected/1bb0507a-76c1-4a53-971c-dce960b71242-kube-api-access-nmmnp\") pod \"test-operator-controller-manager-5854674fcc-tbth4\" (UID: \"1bb0507a-76c1-4a53-971c-dce960b71242\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-tbth4" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.884516 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqvcn\" (UniqueName: \"kubernetes.io/projected/06627027-315a-482d-9e26-32c5b4df115a-kube-api-access-zqvcn\") pod \"telemetry-operator-controller-manager-76cc84c6bb-wv65n\" (UID: \"06627027-315a-482d-9e26-32c5b4df115a\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-wv65n" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.884713 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bvbq\" (UniqueName: \"kubernetes.io/projected/861aea23-001f-4259-b69a-64b95db49c73-kube-api-access-6bvbq\") pod \"placement-operator-controller-manager-78f8948974-4p69t\" (UID: \"861aea23-001f-4259-b69a-64b95db49c73\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-4p69t" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.884964 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zj6h\" (UniqueName: \"kubernetes.io/projected/888d5a2b-8a13-4042-8c43-59241cfdc088-kube-api-access-8zj6h\") pod \"ovn-operator-controller-manager-b6456fdb6-qfp99\" (UID: \"888d5a2b-8a13-4042-8c43-59241cfdc088\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-qfp99" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.885077 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdjm6\" (UniqueName: \"kubernetes.io/projected/ac90f9ba-d2ef-44a1-84e7-8e809b5eef13-kube-api-access-wdjm6\") pod \"swift-operator-controller-manager-5f8c65bbfc-4xvn4\" (UID: \"ac90f9ba-d2ef-44a1-84e7-8e809b5eef13\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-4xvn4" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.917490 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdjm6\" (UniqueName: \"kubernetes.io/projected/ac90f9ba-d2ef-44a1-84e7-8e809b5eef13-kube-api-access-wdjm6\") pod \"swift-operator-controller-manager-5f8c65bbfc-4xvn4\" (UID: \"ac90f9ba-d2ef-44a1-84e7-8e809b5eef13\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-4xvn4" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.931156 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bvbq\" (UniqueName: \"kubernetes.io/projected/861aea23-001f-4259-b69a-64b95db49c73-kube-api-access-6bvbq\") pod \"placement-operator-controller-manager-78f8948974-4p69t\" (UID: \"861aea23-001f-4259-b69a-64b95db49c73\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-4p69t" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.933325 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-576b8587b8-cftqf"] Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.934825 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-576b8587b8-cftqf" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.939181 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.939321 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.939470 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-nkvn4" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.939884 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-576b8587b8-cftqf"] Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.940410 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zj6h\" (UniqueName: \"kubernetes.io/projected/888d5a2b-8a13-4042-8c43-59241cfdc088-kube-api-access-8zj6h\") pod \"ovn-operator-controller-manager-b6456fdb6-qfp99\" (UID: \"888d5a2b-8a13-4042-8c43-59241cfdc088\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-qfp99" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.954960 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-4p69t" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.962777 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fsl4h"] Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.964477 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fsl4h" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.965564 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fsl4h"] Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.969923 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-cql52" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.977003 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-6vnhr"] Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.987705 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74rld\" (UniqueName: \"kubernetes.io/projected/1edee957-6d50-48c2-bccc-e6702bcb5ce0-kube-api-access-74rld\") pod \"watcher-operator-controller-manager-769dc69bc-v4kv7\" (UID: \"1edee957-6d50-48c2-bccc-e6702bcb5ce0\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-v4kv7" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.988493 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8c9ec062-c2b9-47a6-8d7a-eb9f00e63ef3-cert\") pod \"infra-operator-controller-manager-57548d458d-pn7wx\" (UID: \"8c9ec062-c2b9-47a6-8d7a-eb9f00e63ef3\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-pn7wx" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.988536 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmmnp\" (UniqueName: \"kubernetes.io/projected/1bb0507a-76c1-4a53-971c-dce960b71242-kube-api-access-nmmnp\") pod \"test-operator-controller-manager-5854674fcc-tbth4\" (UID: \"1bb0507a-76c1-4a53-971c-dce960b71242\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-tbth4" Nov 29 04:51:40 crc kubenswrapper[4799]: I1129 04:51:40.988598 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqvcn\" (UniqueName: \"kubernetes.io/projected/06627027-315a-482d-9e26-32c5b4df115a-kube-api-access-zqvcn\") pod \"telemetry-operator-controller-manager-76cc84c6bb-wv65n\" (UID: \"06627027-315a-482d-9e26-32c5b4df115a\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-wv65n" Nov 29 04:51:40 crc kubenswrapper[4799]: E1129 04:51:40.988915 4799 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 29 04:51:40 crc kubenswrapper[4799]: E1129 04:51:40.988971 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8c9ec062-c2b9-47a6-8d7a-eb9f00e63ef3-cert podName:8c9ec062-c2b9-47a6-8d7a-eb9f00e63ef3 nodeName:}" failed. No retries permitted until 2025-11-29 04:51:41.988953344 +0000 UTC m=+777.631883744 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/8c9ec062-c2b9-47a6-8d7a-eb9f00e63ef3-cert") pod "infra-operator-controller-manager-57548d458d-pn7wx" (UID: "8c9ec062-c2b9-47a6-8d7a-eb9f00e63ef3") : secret "infra-operator-webhook-server-cert" not found Nov 29 04:51:41 crc kubenswrapper[4799]: I1129 04:51:41.001631 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-4xvn4" Nov 29 04:51:41 crc kubenswrapper[4799]: I1129 04:51:41.008374 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmmnp\" (UniqueName: \"kubernetes.io/projected/1bb0507a-76c1-4a53-971c-dce960b71242-kube-api-access-nmmnp\") pod \"test-operator-controller-manager-5854674fcc-tbth4\" (UID: \"1bb0507a-76c1-4a53-971c-dce960b71242\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-tbth4" Nov 29 04:51:41 crc kubenswrapper[4799]: I1129 04:51:41.031440 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqvcn\" (UniqueName: \"kubernetes.io/projected/06627027-315a-482d-9e26-32c5b4df115a-kube-api-access-zqvcn\") pod \"telemetry-operator-controller-manager-76cc84c6bb-wv65n\" (UID: \"06627027-315a-482d-9e26-32c5b4df115a\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-wv65n" Nov 29 04:51:41 crc kubenswrapper[4799]: I1129 04:51:41.074286 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-wv65n" Nov 29 04:51:41 crc kubenswrapper[4799]: I1129 04:51:41.090678 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/41eeb1eb-9b97-41fa-ad22-09d4d4b34abc-metrics-certs\") pod \"openstack-operator-controller-manager-576b8587b8-cftqf\" (UID: \"41eeb1eb-9b97-41fa-ad22-09d4d4b34abc\") " pod="openstack-operators/openstack-operator-controller-manager-576b8587b8-cftqf" Nov 29 04:51:41 crc kubenswrapper[4799]: I1129 04:51:41.090759 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74rld\" (UniqueName: \"kubernetes.io/projected/1edee957-6d50-48c2-bccc-e6702bcb5ce0-kube-api-access-74rld\") pod \"watcher-operator-controller-manager-769dc69bc-v4kv7\" (UID: \"1edee957-6d50-48c2-bccc-e6702bcb5ce0\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-v4kv7" Nov 29 04:51:41 crc kubenswrapper[4799]: I1129 04:51:41.090925 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xw4st\" (UniqueName: \"kubernetes.io/projected/41eeb1eb-9b97-41fa-ad22-09d4d4b34abc-kube-api-access-xw4st\") pod \"openstack-operator-controller-manager-576b8587b8-cftqf\" (UID: \"41eeb1eb-9b97-41fa-ad22-09d4d4b34abc\") " pod="openstack-operators/openstack-operator-controller-manager-576b8587b8-cftqf" Nov 29 04:51:41 crc kubenswrapper[4799]: I1129 04:51:41.090970 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/41eeb1eb-9b97-41fa-ad22-09d4d4b34abc-webhook-certs\") pod \"openstack-operator-controller-manager-576b8587b8-cftqf\" (UID: \"41eeb1eb-9b97-41fa-ad22-09d4d4b34abc\") " pod="openstack-operators/openstack-operator-controller-manager-576b8587b8-cftqf" Nov 29 04:51:41 crc kubenswrapper[4799]: I1129 04:51:41.090993 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5vm4\" (UniqueName: \"kubernetes.io/projected/225a8e52-78ae-41eb-9c27-64edb57a135e-kube-api-access-z5vm4\") pod \"rabbitmq-cluster-operator-manager-668c99d594-fsl4h\" (UID: \"225a8e52-78ae-41eb-9c27-64edb57a135e\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fsl4h" Nov 29 04:51:41 crc kubenswrapper[4799]: I1129 04:51:41.104566 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-tbth4" Nov 29 04:51:41 crc kubenswrapper[4799]: I1129 04:51:41.142618 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74rld\" (UniqueName: \"kubernetes.io/projected/1edee957-6d50-48c2-bccc-e6702bcb5ce0-kube-api-access-74rld\") pod \"watcher-operator-controller-manager-769dc69bc-v4kv7\" (UID: \"1edee957-6d50-48c2-bccc-e6702bcb5ce0\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-v4kv7" Nov 29 04:51:41 crc kubenswrapper[4799]: I1129 04:51:41.147556 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-668d9c48b9-lkbfl"] Nov 29 04:51:41 crc kubenswrapper[4799]: I1129 04:51:41.192118 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5vm4\" (UniqueName: \"kubernetes.io/projected/225a8e52-78ae-41eb-9c27-64edb57a135e-kube-api-access-z5vm4\") pod \"rabbitmq-cluster-operator-manager-668c99d594-fsl4h\" (UID: \"225a8e52-78ae-41eb-9c27-64edb57a135e\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fsl4h" Nov 29 04:51:41 crc kubenswrapper[4799]: I1129 04:51:41.192206 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/41eeb1eb-9b97-41fa-ad22-09d4d4b34abc-metrics-certs\") pod \"openstack-operator-controller-manager-576b8587b8-cftqf\" (UID: \"41eeb1eb-9b97-41fa-ad22-09d4d4b34abc\") " pod="openstack-operators/openstack-operator-controller-manager-576b8587b8-cftqf" Nov 29 04:51:41 crc kubenswrapper[4799]: I1129 04:51:41.192277 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xw4st\" (UniqueName: \"kubernetes.io/projected/41eeb1eb-9b97-41fa-ad22-09d4d4b34abc-kube-api-access-xw4st\") pod \"openstack-operator-controller-manager-576b8587b8-cftqf\" (UID: \"41eeb1eb-9b97-41fa-ad22-09d4d4b34abc\") " pod="openstack-operators/openstack-operator-controller-manager-576b8587b8-cftqf" Nov 29 04:51:41 crc kubenswrapper[4799]: I1129 04:51:41.192317 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/41eeb1eb-9b97-41fa-ad22-09d4d4b34abc-webhook-certs\") pod \"openstack-operator-controller-manager-576b8587b8-cftqf\" (UID: \"41eeb1eb-9b97-41fa-ad22-09d4d4b34abc\") " pod="openstack-operators/openstack-operator-controller-manager-576b8587b8-cftqf" Nov 29 04:51:41 crc kubenswrapper[4799]: E1129 04:51:41.192466 4799 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 29 04:51:41 crc kubenswrapper[4799]: E1129 04:51:41.192523 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/41eeb1eb-9b97-41fa-ad22-09d4d4b34abc-webhook-certs podName:41eeb1eb-9b97-41fa-ad22-09d4d4b34abc nodeName:}" failed. No retries permitted until 2025-11-29 04:51:41.692508949 +0000 UTC m=+777.335439349 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/41eeb1eb-9b97-41fa-ad22-09d4d4b34abc-webhook-certs") pod "openstack-operator-controller-manager-576b8587b8-cftqf" (UID: "41eeb1eb-9b97-41fa-ad22-09d4d4b34abc") : secret "webhook-server-cert" not found Nov 29 04:51:41 crc kubenswrapper[4799]: E1129 04:51:41.192687 4799 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 29 04:51:41 crc kubenswrapper[4799]: E1129 04:51:41.192712 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/41eeb1eb-9b97-41fa-ad22-09d4d4b34abc-metrics-certs podName:41eeb1eb-9b97-41fa-ad22-09d4d4b34abc nodeName:}" failed. No retries permitted until 2025-11-29 04:51:41.692704993 +0000 UTC m=+777.335635393 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/41eeb1eb-9b97-41fa-ad22-09d4d4b34abc-metrics-certs") pod "openstack-operator-controller-manager-576b8587b8-cftqf" (UID: "41eeb1eb-9b97-41fa-ad22-09d4d4b34abc") : secret "metrics-server-cert" not found Nov 29 04:51:41 crc kubenswrapper[4799]: I1129 04:51:41.210114 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xw4st\" (UniqueName: \"kubernetes.io/projected/41eeb1eb-9b97-41fa-ad22-09d4d4b34abc-kube-api-access-xw4st\") pod \"openstack-operator-controller-manager-576b8587b8-cftqf\" (UID: \"41eeb1eb-9b97-41fa-ad22-09d4d4b34abc\") " pod="openstack-operators/openstack-operator-controller-manager-576b8587b8-cftqf" Nov 29 04:51:41 crc kubenswrapper[4799]: I1129 04:51:41.211373 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5vm4\" (UniqueName: \"kubernetes.io/projected/225a8e52-78ae-41eb-9c27-64edb57a135e-kube-api-access-z5vm4\") pod \"rabbitmq-cluster-operator-manager-668c99d594-fsl4h\" (UID: \"225a8e52-78ae-41eb-9c27-64edb57a135e\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fsl4h" Nov 29 04:51:41 crc kubenswrapper[4799]: W1129 04:51:41.211523 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod743ba96f_315f_4e0d_abb0_5a6389140c7b.slice/crio-b1f2e1778073d87468527153ee3ad83d9985b882ab635a55b2cfca1324015efe WatchSource:0}: Error finding container b1f2e1778073d87468527153ee3ad83d9985b882ab635a55b2cfca1324015efe: Status 404 returned error can't find the container with id b1f2e1778073d87468527153ee3ad83d9985b882ab635a55b2cfca1324015efe Nov 29 04:51:41 crc kubenswrapper[4799]: I1129 04:51:41.213305 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-v4kv7" Nov 29 04:51:41 crc kubenswrapper[4799]: I1129 04:51:41.232881 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-qfp99" Nov 29 04:51:41 crc kubenswrapper[4799]: I1129 04:51:41.293431 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8e0677fe-7f53-4e50-881b-943fffe1c0ff-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4qnwnt\" (UID: \"8e0677fe-7f53-4e50-881b-943fffe1c0ff\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4qnwnt" Nov 29 04:51:41 crc kubenswrapper[4799]: E1129 04:51:41.293749 4799 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 29 04:51:41 crc kubenswrapper[4799]: E1129 04:51:41.293885 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8e0677fe-7f53-4e50-881b-943fffe1c0ff-cert podName:8e0677fe-7f53-4e50-881b-943fffe1c0ff nodeName:}" failed. No retries permitted until 2025-11-29 04:51:42.293861115 +0000 UTC m=+777.936791515 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/8e0677fe-7f53-4e50-881b-943fffe1c0ff-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4qnwnt" (UID: "8e0677fe-7f53-4e50-881b-943fffe1c0ff") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 29 04:51:41 crc kubenswrapper[4799]: I1129 04:51:41.329756 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fsl4h" Nov 29 04:51:41 crc kubenswrapper[4799]: I1129 04:51:41.615882 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kxgw8"] Nov 29 04:51:41 crc kubenswrapper[4799]: W1129 04:51:41.632296 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod94ec7bc8_8413_4040_90dd_a5c9ce4ca98f.slice/crio-9d3a81482d6c7231790df8bc4b78c5a7d624ff70b2e3dd1c42f45dea526e0646 WatchSource:0}: Error finding container 9d3a81482d6c7231790df8bc4b78c5a7d624ff70b2e3dd1c42f45dea526e0646: Status 404 returned error can't find the container with id 9d3a81482d6c7231790df8bc4b78c5a7d624ff70b2e3dd1c42f45dea526e0646 Nov 29 04:51:41 crc kubenswrapper[4799]: I1129 04:51:41.646061 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-84f754f96-t5gjl"] Nov 29 04:51:41 crc kubenswrapper[4799]: W1129 04:51:41.660702 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4bf2a041_0f89_480e_88d2_ba630db189c1.slice/crio-ebb8d6e94c1bfba354fe7a2c7d6fd69567953c68d079f785999dfb3fb5787dc2 WatchSource:0}: Error finding container ebb8d6e94c1bfba354fe7a2c7d6fd69567953c68d079f785999dfb3fb5787dc2: Status 404 returned error can't find the container with id ebb8d6e94c1bfba354fe7a2c7d6fd69567953c68d079f785999dfb3fb5787dc2 Nov 29 04:51:41 crc kubenswrapper[4799]: I1129 04:51:41.676412 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-p555g"] Nov 29 04:51:41 crc kubenswrapper[4799]: I1129 04:51:41.684589 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-546d4bdf48-8p2z8"] Nov 29 04:51:41 crc kubenswrapper[4799]: I1129 04:51:41.698715 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/41eeb1eb-9b97-41fa-ad22-09d4d4b34abc-metrics-certs\") pod \"openstack-operator-controller-manager-576b8587b8-cftqf\" (UID: \"41eeb1eb-9b97-41fa-ad22-09d4d4b34abc\") " pod="openstack-operators/openstack-operator-controller-manager-576b8587b8-cftqf" Nov 29 04:51:41 crc kubenswrapper[4799]: I1129 04:51:41.699112 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/41eeb1eb-9b97-41fa-ad22-09d4d4b34abc-webhook-certs\") pod \"openstack-operator-controller-manager-576b8587b8-cftqf\" (UID: \"41eeb1eb-9b97-41fa-ad22-09d4d4b34abc\") " pod="openstack-operators/openstack-operator-controller-manager-576b8587b8-cftqf" Nov 29 04:51:41 crc kubenswrapper[4799]: E1129 04:51:41.699187 4799 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 29 04:51:41 crc kubenswrapper[4799]: E1129 04:51:41.699288 4799 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 29 04:51:41 crc kubenswrapper[4799]: E1129 04:51:41.699301 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/41eeb1eb-9b97-41fa-ad22-09d4d4b34abc-metrics-certs podName:41eeb1eb-9b97-41fa-ad22-09d4d4b34abc nodeName:}" failed. No retries permitted until 2025-11-29 04:51:42.699274454 +0000 UTC m=+778.342204914 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/41eeb1eb-9b97-41fa-ad22-09d4d4b34abc-metrics-certs") pod "openstack-operator-controller-manager-576b8587b8-cftqf" (UID: "41eeb1eb-9b97-41fa-ad22-09d4d4b34abc") : secret "metrics-server-cert" not found Nov 29 04:51:41 crc kubenswrapper[4799]: E1129 04:51:41.699407 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/41eeb1eb-9b97-41fa-ad22-09d4d4b34abc-webhook-certs podName:41eeb1eb-9b97-41fa-ad22-09d4d4b34abc nodeName:}" failed. No retries permitted until 2025-11-29 04:51:42.699379937 +0000 UTC m=+778.342310347 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/41eeb1eb-9b97-41fa-ad22-09d4d4b34abc-webhook-certs") pod "openstack-operator-controller-manager-576b8587b8-cftqf" (UID: "41eeb1eb-9b97-41fa-ad22-09d4d4b34abc") : secret "webhook-server-cert" not found Nov 29 04:51:41 crc kubenswrapper[4799]: I1129 04:51:41.743081 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kxgw8" event={"ID":"94ec7bc8-8413-4040-90dd-a5c9ce4ca98f","Type":"ContainerStarted","Data":"9d3a81482d6c7231790df8bc4b78c5a7d624ff70b2e3dd1c42f45dea526e0646"} Nov 29 04:51:41 crc kubenswrapper[4799]: I1129 04:51:41.745080 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-8p2z8" event={"ID":"f696db58-bd41-4db0-9ec7-c4a4710e6b8e","Type":"ContainerStarted","Data":"a047b5767bbb17171202338b9b065f7a09f2cc3b572ffeb317f5ee5c30394e37"} Nov 29 04:51:41 crc kubenswrapper[4799]: I1129 04:51:41.761377 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-6vnhr" event={"ID":"7232eadb-c887-440b-b6f9-bffa4a6dc871","Type":"ContainerStarted","Data":"a01238e9f4e758d060b4e35d1327ca3dbd27a78b3eef5e5506da6095c0608183"} Nov 29 04:51:41 crc kubenswrapper[4799]: I1129 04:51:41.765466 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-84f754f96-t5gjl" event={"ID":"4bf2a041-0f89-480e-88d2-ba630db189c1","Type":"ContainerStarted","Data":"ebb8d6e94c1bfba354fe7a2c7d6fd69567953c68d079f785999dfb3fb5787dc2"} Nov 29 04:51:41 crc kubenswrapper[4799]: I1129 04:51:41.766835 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-lkbfl" event={"ID":"743ba96f-315f-4e0d-abb0-5a6389140c7b","Type":"ContainerStarted","Data":"b1f2e1778073d87468527153ee3ad83d9985b882ab635a55b2cfca1324015efe"} Nov 29 04:51:41 crc kubenswrapper[4799]: I1129 04:51:41.768283 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-p555g" event={"ID":"2798aeab-f9b9-4728-ad63-186b16afcc31","Type":"ContainerStarted","Data":"59afd84c5627fd5f00957995c960617c566b4427c5515674801f4288a5609a6a"} Nov 29 04:51:41 crc kubenswrapper[4799]: I1129 04:51:41.972018 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-sfpmg"] Nov 29 04:51:41 crc kubenswrapper[4799]: I1129 04:51:41.982575 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-pc2zl"] Nov 29 04:51:41 crc kubenswrapper[4799]: I1129 04:51:41.989914 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-25fzb"] Nov 29 04:51:41 crc kubenswrapper[4799]: I1129 04:51:41.998369 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-7zddj"] Nov 29 04:51:42 crc kubenswrapper[4799]: W1129 04:51:42.002666 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podff72829c_dfd8_4ae2_b508_f89e40f654eb.slice/crio-afa38ee450461e5e26db9c07d3ce88cd3b67342889ef9338cf06cfb72e1fd483 WatchSource:0}: Error finding container afa38ee450461e5e26db9c07d3ce88cd3b67342889ef9338cf06cfb72e1fd483: Status 404 returned error can't find the container with id afa38ee450461e5e26db9c07d3ce88cd3b67342889ef9338cf06cfb72e1fd483 Nov 29 04:51:42 crc kubenswrapper[4799]: I1129 04:51:42.006995 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8c9ec062-c2b9-47a6-8d7a-eb9f00e63ef3-cert\") pod \"infra-operator-controller-manager-57548d458d-pn7wx\" (UID: \"8c9ec062-c2b9-47a6-8d7a-eb9f00e63ef3\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-pn7wx" Nov 29 04:51:42 crc kubenswrapper[4799]: E1129 04:51:42.007196 4799 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 29 04:51:42 crc kubenswrapper[4799]: E1129 04:51:42.007270 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8c9ec062-c2b9-47a6-8d7a-eb9f00e63ef3-cert podName:8c9ec062-c2b9-47a6-8d7a-eb9f00e63ef3 nodeName:}" failed. No retries permitted until 2025-11-29 04:51:44.007251477 +0000 UTC m=+779.650181887 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/8c9ec062-c2b9-47a6-8d7a-eb9f00e63ef3-cert") pod "infra-operator-controller-manager-57548d458d-pn7wx" (UID: "8c9ec062-c2b9-47a6-8d7a-eb9f00e63ef3") : secret "infra-operator-webhook-server-cert" not found Nov 29 04:51:42 crc kubenswrapper[4799]: I1129 04:51:42.209054 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-8nl87"] Nov 29 04:51:42 crc kubenswrapper[4799]: W1129 04:51:42.211100 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbdcce387_cd82_490e_bd50_cf47158b1ea6.slice/crio-205ca2ebd1d3e2c4c155d8ce3bc2ce1bfa612584a10e866ae95d188268c6a32f WatchSource:0}: Error finding container 205ca2ebd1d3e2c4c155d8ce3bc2ce1bfa612584a10e866ae95d188268c6a32f: Status 404 returned error can't find the container with id 205ca2ebd1d3e2c4c155d8ce3bc2ce1bfa612584a10e866ae95d188268c6a32f Nov 29 04:51:42 crc kubenswrapper[4799]: I1129 04:51:42.223372 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-wv65n"] Nov 29 04:51:42 crc kubenswrapper[4799]: I1129 04:51:42.226669 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-4p69t"] Nov 29 04:51:42 crc kubenswrapper[4799]: W1129 04:51:42.236573 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod888d5a2b_8a13_4042_8c43_59241cfdc088.slice/crio-3aaef093581b898d9884a1c35b1ed1c51138933d95610acf249217750806674a WatchSource:0}: Error finding container 3aaef093581b898d9884a1c35b1ed1c51138933d95610acf249217750806674a: Status 404 returned error can't find the container with id 3aaef093581b898d9884a1c35b1ed1c51138933d95610acf249217750806674a Nov 29 04:51:42 crc kubenswrapper[4799]: I1129 04:51:42.237865 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-pkrhh"] Nov 29 04:51:42 crc kubenswrapper[4799]: I1129 04:51:42.245223 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-wstbs"] Nov 29 04:51:42 crc kubenswrapper[4799]: W1129 04:51:42.252373 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod225a8e52_78ae_41eb_9c27_64edb57a135e.slice/crio-f014f82b045020e2b334abde1c7b3223cb4ed4aafbdf22ba5f18f7964c55a588 WatchSource:0}: Error finding container f014f82b045020e2b334abde1c7b3223cb4ed4aafbdf22ba5f18f7964c55a588: Status 404 returned error can't find the container with id f014f82b045020e2b334abde1c7b3223cb4ed4aafbdf22ba5f18f7964c55a588 Nov 29 04:51:42 crc kubenswrapper[4799]: I1129 04:51:42.253364 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fsl4h"] Nov 29 04:51:42 crc kubenswrapper[4799]: I1129 04:51:42.260461 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-qfp99"] Nov 29 04:51:42 crc kubenswrapper[4799]: E1129 04:51:42.264934 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/designate-operator@sha256:9f68d7bc8c6bce38f46dee8a8272d5365c49fe7b32b2af52e8ac884e212f3a85,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9jq9b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-78b4bc895b-pkrhh_openstack-operators(89038368-29a3-417e-8ffa-819e5901a52a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 29 04:51:42 crc kubenswrapper[4799]: E1129 04:51:42.265060 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6bvbq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-4p69t_openstack-operators(861aea23-001f-4259-b69a-64b95db49c73): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 29 04:51:42 crc kubenswrapper[4799]: I1129 04:51:42.267208 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-4xvn4"] Nov 29 04:51:42 crc kubenswrapper[4799]: E1129 04:51:42.268513 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6bvbq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-4p69t_openstack-operators(861aea23-001f-4259-b69a-64b95db49c73): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 29 04:51:42 crc kubenswrapper[4799]: E1129 04:51:42.268542 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9jq9b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-78b4bc895b-pkrhh_openstack-operators(89038368-29a3-417e-8ffa-819e5901a52a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 29 04:51:42 crc kubenswrapper[4799]: E1129 04:51:42.269632 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-4p69t" podUID="861aea23-001f-4259-b69a-64b95db49c73" Nov 29 04:51:42 crc kubenswrapper[4799]: E1129 04:51:42.269669 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-pkrhh" podUID="89038368-29a3-417e-8ffa-819e5901a52a" Nov 29 04:51:42 crc kubenswrapper[4799]: E1129 04:51:42.270735 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nmmnp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-tbth4_openstack-operators(1bb0507a-76c1-4a53-971c-dce960b71242): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 29 04:51:42 crc kubenswrapper[4799]: I1129 04:51:42.271368 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-v4kv7"] Nov 29 04:51:42 crc kubenswrapper[4799]: E1129 04:51:42.273327 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nmmnp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-tbth4_openstack-operators(1bb0507a-76c1-4a53-971c-dce960b71242): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 29 04:51:42 crc kubenswrapper[4799]: E1129 04:51:42.274515 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-tbth4" podUID="1bb0507a-76c1-4a53-971c-dce960b71242" Nov 29 04:51:42 crc kubenswrapper[4799]: I1129 04:51:42.275455 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-tbth4"] Nov 29 04:51:42 crc kubenswrapper[4799]: W1129 04:51:42.277487 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac90f9ba_d2ef_44a1_84e7_8e809b5eef13.slice/crio-247112dcc3f01ad87b0225b908abfe898929cf2be4af723a54dd745f3ff48372 WatchSource:0}: Error finding container 247112dcc3f01ad87b0225b908abfe898929cf2be4af723a54dd745f3ff48372: Status 404 returned error can't find the container with id 247112dcc3f01ad87b0225b908abfe898929cf2be4af723a54dd745f3ff48372 Nov 29 04:51:42 crc kubenswrapper[4799]: W1129 04:51:42.278697 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1edee957_6d50_48c2_bccc_e6702bcb5ce0.slice/crio-ba2df6bfc78c27fcc338f3a8b47673656c70ff1e921f33e805380764ee1a263f WatchSource:0}: Error finding container ba2df6bfc78c27fcc338f3a8b47673656c70ff1e921f33e805380764ee1a263f: Status 404 returned error can't find the container with id ba2df6bfc78c27fcc338f3a8b47673656c70ff1e921f33e805380764ee1a263f Nov 29 04:51:42 crc kubenswrapper[4799]: E1129 04:51:42.281713 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wdjm6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-4xvn4_openstack-operators(ac90f9ba-d2ef-44a1-84e7-8e809b5eef13): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 29 04:51:42 crc kubenswrapper[4799]: E1129 04:51:42.282411 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-74rld,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-v4kv7_openstack-operators(1edee957-6d50-48c2-bccc-e6702bcb5ce0): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 29 04:51:42 crc kubenswrapper[4799]: E1129 04:51:42.285147 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wdjm6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-4xvn4_openstack-operators(ac90f9ba-d2ef-44a1-84e7-8e809b5eef13): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 29 04:51:42 crc kubenswrapper[4799]: E1129 04:51:42.285334 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-74rld,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-v4kv7_openstack-operators(1edee957-6d50-48c2-bccc-e6702bcb5ce0): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 29 04:51:42 crc kubenswrapper[4799]: E1129 04:51:42.286852 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-v4kv7" podUID="1edee957-6d50-48c2-bccc-e6702bcb5ce0" Nov 29 04:51:42 crc kubenswrapper[4799]: E1129 04:51:42.286874 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-4xvn4" podUID="ac90f9ba-d2ef-44a1-84e7-8e809b5eef13" Nov 29 04:51:42 crc kubenswrapper[4799]: I1129 04:51:42.316390 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8e0677fe-7f53-4e50-881b-943fffe1c0ff-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4qnwnt\" (UID: \"8e0677fe-7f53-4e50-881b-943fffe1c0ff\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4qnwnt" Nov 29 04:51:42 crc kubenswrapper[4799]: E1129 04:51:42.316829 4799 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 29 04:51:42 crc kubenswrapper[4799]: E1129 04:51:42.316907 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8e0677fe-7f53-4e50-881b-943fffe1c0ff-cert podName:8e0677fe-7f53-4e50-881b-943fffe1c0ff nodeName:}" failed. No retries permitted until 2025-11-29 04:51:44.316888618 +0000 UTC m=+779.959819028 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/8e0677fe-7f53-4e50-881b-943fffe1c0ff-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4qnwnt" (UID: "8e0677fe-7f53-4e50-881b-943fffe1c0ff") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 29 04:51:42 crc kubenswrapper[4799]: I1129 04:51:42.723613 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/41eeb1eb-9b97-41fa-ad22-09d4d4b34abc-webhook-certs\") pod \"openstack-operator-controller-manager-576b8587b8-cftqf\" (UID: \"41eeb1eb-9b97-41fa-ad22-09d4d4b34abc\") " pod="openstack-operators/openstack-operator-controller-manager-576b8587b8-cftqf" Nov 29 04:51:42 crc kubenswrapper[4799]: I1129 04:51:42.723691 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/41eeb1eb-9b97-41fa-ad22-09d4d4b34abc-metrics-certs\") pod \"openstack-operator-controller-manager-576b8587b8-cftqf\" (UID: \"41eeb1eb-9b97-41fa-ad22-09d4d4b34abc\") " pod="openstack-operators/openstack-operator-controller-manager-576b8587b8-cftqf" Nov 29 04:51:42 crc kubenswrapper[4799]: E1129 04:51:42.723873 4799 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 29 04:51:42 crc kubenswrapper[4799]: E1129 04:51:42.723925 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/41eeb1eb-9b97-41fa-ad22-09d4d4b34abc-metrics-certs podName:41eeb1eb-9b97-41fa-ad22-09d4d4b34abc nodeName:}" failed. No retries permitted until 2025-11-29 04:51:44.723909883 +0000 UTC m=+780.366840283 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/41eeb1eb-9b97-41fa-ad22-09d4d4b34abc-metrics-certs") pod "openstack-operator-controller-manager-576b8587b8-cftqf" (UID: "41eeb1eb-9b97-41fa-ad22-09d4d4b34abc") : secret "metrics-server-cert" not found Nov 29 04:51:42 crc kubenswrapper[4799]: E1129 04:51:42.723947 4799 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 29 04:51:42 crc kubenswrapper[4799]: E1129 04:51:42.723981 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/41eeb1eb-9b97-41fa-ad22-09d4d4b34abc-webhook-certs podName:41eeb1eb-9b97-41fa-ad22-09d4d4b34abc nodeName:}" failed. No retries permitted until 2025-11-29 04:51:44.723971185 +0000 UTC m=+780.366901585 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/41eeb1eb-9b97-41fa-ad22-09d4d4b34abc-webhook-certs") pod "openstack-operator-controller-manager-576b8587b8-cftqf" (UID: "41eeb1eb-9b97-41fa-ad22-09d4d4b34abc") : secret "webhook-server-cert" not found Nov 29 04:51:42 crc kubenswrapper[4799]: I1129 04:51:42.805617 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fsl4h" event={"ID":"225a8e52-78ae-41eb-9c27-64edb57a135e","Type":"ContainerStarted","Data":"f014f82b045020e2b334abde1c7b3223cb4ed4aafbdf22ba5f18f7964c55a588"} Nov 29 04:51:42 crc kubenswrapper[4799]: I1129 04:51:42.812661 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-tbth4" event={"ID":"1bb0507a-76c1-4a53-971c-dce960b71242","Type":"ContainerStarted","Data":"cd9314cc7138c315101779ce59755f52f99f4e018e86e73931d6ec7c9c0719c8"} Nov 29 04:51:42 crc kubenswrapper[4799]: I1129 04:51:42.814958 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-pc2zl" event={"ID":"e1d949de-4ce7-4016-83fd-4c0ce003eb0f","Type":"ContainerStarted","Data":"bd0bb64d0d7b5b31eb8208f0b033809d0f1651845d8581db13f7361f7e32eed0"} Nov 29 04:51:42 crc kubenswrapper[4799]: E1129 04:51:42.826723 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-tbth4" podUID="1bb0507a-76c1-4a53-971c-dce960b71242" Nov 29 04:51:42 crc kubenswrapper[4799]: I1129 04:51:42.855098 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-sfpmg" event={"ID":"7851df67-88d5-45d8-804e-5d3bf3fdb3af","Type":"ContainerStarted","Data":"9039e88d4c40380907b2b0a4a9454523261f95b74d5ed015910d547eef79b3bc"} Nov 29 04:51:42 crc kubenswrapper[4799]: I1129 04:51:42.858047 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-wstbs" event={"ID":"bbe2e2fb-4619-405c-a9ad-4bdf727882cd","Type":"ContainerStarted","Data":"c8d6c11d4693c890c3f32eb54021181809499978489b622a31eb66441a3ef53e"} Nov 29 04:51:42 crc kubenswrapper[4799]: I1129 04:51:42.865414 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-wv65n" event={"ID":"06627027-315a-482d-9e26-32c5b4df115a","Type":"ContainerStarted","Data":"56548d611da81d83a27c14425cc5a7e3b609316956a7bdff69eaf1e24744f983"} Nov 29 04:51:42 crc kubenswrapper[4799]: I1129 04:51:42.868770 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-25fzb" event={"ID":"ff72829c-dfd8-4ae2-b508-f89e40f654eb","Type":"ContainerStarted","Data":"afa38ee450461e5e26db9c07d3ce88cd3b67342889ef9338cf06cfb72e1fd483"} Nov 29 04:51:42 crc kubenswrapper[4799]: I1129 04:51:42.872136 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-v4kv7" event={"ID":"1edee957-6d50-48c2-bccc-e6702bcb5ce0","Type":"ContainerStarted","Data":"ba2df6bfc78c27fcc338f3a8b47673656c70ff1e921f33e805380764ee1a263f"} Nov 29 04:51:42 crc kubenswrapper[4799]: E1129 04:51:42.875304 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-v4kv7" podUID="1edee957-6d50-48c2-bccc-e6702bcb5ce0" Nov 29 04:51:42 crc kubenswrapper[4799]: I1129 04:51:42.882242 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-4xvn4" event={"ID":"ac90f9ba-d2ef-44a1-84e7-8e809b5eef13","Type":"ContainerStarted","Data":"247112dcc3f01ad87b0225b908abfe898929cf2be4af723a54dd745f3ff48372"} Nov 29 04:51:42 crc kubenswrapper[4799]: E1129 04:51:42.885575 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-4xvn4" podUID="ac90f9ba-d2ef-44a1-84e7-8e809b5eef13" Nov 29 04:51:42 crc kubenswrapper[4799]: I1129 04:51:42.885585 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-pkrhh" event={"ID":"89038368-29a3-417e-8ffa-819e5901a52a","Type":"ContainerStarted","Data":"029a7030cc4651a9b85007d4828f9d994e445b2e0095359aa7363b487c14bb1b"} Nov 29 04:51:42 crc kubenswrapper[4799]: E1129 04:51:42.896711 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/designate-operator@sha256:9f68d7bc8c6bce38f46dee8a8272d5365c49fe7b32b2af52e8ac884e212f3a85\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-pkrhh" podUID="89038368-29a3-417e-8ffa-819e5901a52a" Nov 29 04:51:42 crc kubenswrapper[4799]: I1129 04:51:42.909063 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-4p69t" event={"ID":"861aea23-001f-4259-b69a-64b95db49c73","Type":"ContainerStarted","Data":"039b5c82418e4d03fdbbd25df04b48e115fa17069a9c2e1b8643523cc1319655"} Nov 29 04:51:42 crc kubenswrapper[4799]: I1129 04:51:42.911557 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-qfp99" event={"ID":"888d5a2b-8a13-4042-8c43-59241cfdc088","Type":"ContainerStarted","Data":"3aaef093581b898d9884a1c35b1ed1c51138933d95610acf249217750806674a"} Nov 29 04:51:42 crc kubenswrapper[4799]: I1129 04:51:42.913109 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-8nl87" event={"ID":"bdcce387-cd82-490e-bd50-cf47158b1ea6","Type":"ContainerStarted","Data":"205ca2ebd1d3e2c4c155d8ce3bc2ce1bfa612584a10e866ae95d188268c6a32f"} Nov 29 04:51:42 crc kubenswrapper[4799]: E1129 04:51:42.913431 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-4p69t" podUID="861aea23-001f-4259-b69a-64b95db49c73" Nov 29 04:51:42 crc kubenswrapper[4799]: I1129 04:51:42.915309 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-7zddj" event={"ID":"259abc1e-a30c-4a8b-acd1-18052862198e","Type":"ContainerStarted","Data":"76d9725b97193d97e8ae576c9ecadfbe750ce269799f38edf484145c9a2f29fc"} Nov 29 04:51:43 crc kubenswrapper[4799]: E1129 04:51:43.621034 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = parsing image configuration: Get \"https://cdn01.quay.io/quayio-production-s3/sha256/2e/2e965a08b1e129438034a39257c39c4f41bcefa3d8de39716f1e53ad90f967f2?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIATAAF2YHTGR23ZTE6%2F20251129%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20251129T045142Z&X-Amz-Expires=600&X-Amz-SignedHeaders=host&X-Amz-Signature=ba292f783ddde29d42f39baef54928665bba33bb8942cd267e625d402b5f669d®ion=us-east-1&namespace=openstack-k8s-operators&username=openshift-release-dev+ocm_access_1b89217552bc42d1be3fb06a1aed001a&repo_name=neutron-operator&akamai_signature=exp=1764392802~hmac=a85ab262adc9a7f756e25ba39129ce6236b33a2cd435f1795733e94cd99ffae1\": remote error: tls: internal error" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557" Nov 29 04:51:43 crc kubenswrapper[4799]: E1129 04:51:43.621677 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bt8cm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-pc2zl_openstack-operators(e1d949de-4ce7-4016-83fd-4c0ce003eb0f): ErrImagePull: parsing image configuration: Get \"https://cdn01.quay.io/quayio-production-s3/sha256/2e/2e965a08b1e129438034a39257c39c4f41bcefa3d8de39716f1e53ad90f967f2?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIATAAF2YHTGR23ZTE6%2F20251129%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20251129T045142Z&X-Amz-Expires=600&X-Amz-SignedHeaders=host&X-Amz-Signature=ba292f783ddde29d42f39baef54928665bba33bb8942cd267e625d402b5f669d®ion=us-east-1&namespace=openstack-k8s-operators&username=openshift-release-dev+ocm_access_1b89217552bc42d1be3fb06a1aed001a&repo_name=neutron-operator&akamai_signature=exp=1764392802~hmac=a85ab262adc9a7f756e25ba39129ce6236b33a2cd435f1795733e94cd99ffae1\": remote error: tls: internal error" logger="UnhandledError" Nov 29 04:51:43 crc kubenswrapper[4799]: E1129 04:51:43.966586 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-4xvn4" podUID="ac90f9ba-d2ef-44a1-84e7-8e809b5eef13" Nov 29 04:51:43 crc kubenswrapper[4799]: E1129 04:51:43.967119 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-v4kv7" podUID="1edee957-6d50-48c2-bccc-e6702bcb5ce0" Nov 29 04:51:43 crc kubenswrapper[4799]: E1129 04:51:43.967316 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/designate-operator@sha256:9f68d7bc8c6bce38f46dee8a8272d5365c49fe7b32b2af52e8ac884e212f3a85\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-pkrhh" podUID="89038368-29a3-417e-8ffa-819e5901a52a" Nov 29 04:51:43 crc kubenswrapper[4799]: E1129 04:51:43.968375 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-4p69t" podUID="861aea23-001f-4259-b69a-64b95db49c73" Nov 29 04:51:43 crc kubenswrapper[4799]: E1129 04:51:43.980522 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-tbth4" podUID="1bb0507a-76c1-4a53-971c-dce960b71242" Nov 29 04:51:44 crc kubenswrapper[4799]: I1129 04:51:44.054324 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8c9ec062-c2b9-47a6-8d7a-eb9f00e63ef3-cert\") pod \"infra-operator-controller-manager-57548d458d-pn7wx\" (UID: \"8c9ec062-c2b9-47a6-8d7a-eb9f00e63ef3\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-pn7wx" Nov 29 04:51:44 crc kubenswrapper[4799]: E1129 04:51:44.054525 4799 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 29 04:51:44 crc kubenswrapper[4799]: E1129 04:51:44.054619 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8c9ec062-c2b9-47a6-8d7a-eb9f00e63ef3-cert podName:8c9ec062-c2b9-47a6-8d7a-eb9f00e63ef3 nodeName:}" failed. No retries permitted until 2025-11-29 04:51:48.054598192 +0000 UTC m=+783.697528592 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/8c9ec062-c2b9-47a6-8d7a-eb9f00e63ef3-cert") pod "infra-operator-controller-manager-57548d458d-pn7wx" (UID: "8c9ec062-c2b9-47a6-8d7a-eb9f00e63ef3") : secret "infra-operator-webhook-server-cert" not found Nov 29 04:51:44 crc kubenswrapper[4799]: I1129 04:51:44.360031 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8e0677fe-7f53-4e50-881b-943fffe1c0ff-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4qnwnt\" (UID: \"8e0677fe-7f53-4e50-881b-943fffe1c0ff\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4qnwnt" Nov 29 04:51:44 crc kubenswrapper[4799]: E1129 04:51:44.360271 4799 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 29 04:51:44 crc kubenswrapper[4799]: E1129 04:51:44.360327 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8e0677fe-7f53-4e50-881b-943fffe1c0ff-cert podName:8e0677fe-7f53-4e50-881b-943fffe1c0ff nodeName:}" failed. No retries permitted until 2025-11-29 04:51:48.360312751 +0000 UTC m=+784.003243151 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/8e0677fe-7f53-4e50-881b-943fffe1c0ff-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4qnwnt" (UID: "8e0677fe-7f53-4e50-881b-943fffe1c0ff") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 29 04:51:44 crc kubenswrapper[4799]: I1129 04:51:44.782891 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/41eeb1eb-9b97-41fa-ad22-09d4d4b34abc-webhook-certs\") pod \"openstack-operator-controller-manager-576b8587b8-cftqf\" (UID: \"41eeb1eb-9b97-41fa-ad22-09d4d4b34abc\") " pod="openstack-operators/openstack-operator-controller-manager-576b8587b8-cftqf" Nov 29 04:51:44 crc kubenswrapper[4799]: I1129 04:51:44.783057 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/41eeb1eb-9b97-41fa-ad22-09d4d4b34abc-metrics-certs\") pod \"openstack-operator-controller-manager-576b8587b8-cftqf\" (UID: \"41eeb1eb-9b97-41fa-ad22-09d4d4b34abc\") " pod="openstack-operators/openstack-operator-controller-manager-576b8587b8-cftqf" Nov 29 04:51:44 crc kubenswrapper[4799]: E1129 04:51:44.783104 4799 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 29 04:51:44 crc kubenswrapper[4799]: E1129 04:51:44.783192 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/41eeb1eb-9b97-41fa-ad22-09d4d4b34abc-webhook-certs podName:41eeb1eb-9b97-41fa-ad22-09d4d4b34abc nodeName:}" failed. No retries permitted until 2025-11-29 04:51:48.783159325 +0000 UTC m=+784.426089725 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/41eeb1eb-9b97-41fa-ad22-09d4d4b34abc-webhook-certs") pod "openstack-operator-controller-manager-576b8587b8-cftqf" (UID: "41eeb1eb-9b97-41fa-ad22-09d4d4b34abc") : secret "webhook-server-cert" not found Nov 29 04:51:44 crc kubenswrapper[4799]: E1129 04:51:44.783305 4799 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 29 04:51:44 crc kubenswrapper[4799]: E1129 04:51:44.783403 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/41eeb1eb-9b97-41fa-ad22-09d4d4b34abc-metrics-certs podName:41eeb1eb-9b97-41fa-ad22-09d4d4b34abc nodeName:}" failed. No retries permitted until 2025-11-29 04:51:48.78338938 +0000 UTC m=+784.426319780 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/41eeb1eb-9b97-41fa-ad22-09d4d4b34abc-metrics-certs") pod "openstack-operator-controller-manager-576b8587b8-cftqf" (UID: "41eeb1eb-9b97-41fa-ad22-09d4d4b34abc") : secret "metrics-server-cert" not found Nov 29 04:51:48 crc kubenswrapper[4799]: I1129 04:51:48.150854 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8c9ec062-c2b9-47a6-8d7a-eb9f00e63ef3-cert\") pod \"infra-operator-controller-manager-57548d458d-pn7wx\" (UID: \"8c9ec062-c2b9-47a6-8d7a-eb9f00e63ef3\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-pn7wx" Nov 29 04:51:48 crc kubenswrapper[4799]: E1129 04:51:48.152007 4799 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 29 04:51:48 crc kubenswrapper[4799]: E1129 04:51:48.152076 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8c9ec062-c2b9-47a6-8d7a-eb9f00e63ef3-cert podName:8c9ec062-c2b9-47a6-8d7a-eb9f00e63ef3 nodeName:}" failed. No retries permitted until 2025-11-29 04:51:56.152054486 +0000 UTC m=+791.794984896 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/8c9ec062-c2b9-47a6-8d7a-eb9f00e63ef3-cert") pod "infra-operator-controller-manager-57548d458d-pn7wx" (UID: "8c9ec062-c2b9-47a6-8d7a-eb9f00e63ef3") : secret "infra-operator-webhook-server-cert" not found Nov 29 04:51:48 crc kubenswrapper[4799]: I1129 04:51:48.457884 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8e0677fe-7f53-4e50-881b-943fffe1c0ff-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4qnwnt\" (UID: \"8e0677fe-7f53-4e50-881b-943fffe1c0ff\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4qnwnt" Nov 29 04:51:48 crc kubenswrapper[4799]: E1129 04:51:48.458069 4799 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 29 04:51:48 crc kubenswrapper[4799]: E1129 04:51:48.458143 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8e0677fe-7f53-4e50-881b-943fffe1c0ff-cert podName:8e0677fe-7f53-4e50-881b-943fffe1c0ff nodeName:}" failed. No retries permitted until 2025-11-29 04:51:56.458120674 +0000 UTC m=+792.101051074 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/8e0677fe-7f53-4e50-881b-943fffe1c0ff-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4qnwnt" (UID: "8e0677fe-7f53-4e50-881b-943fffe1c0ff") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 29 04:51:48 crc kubenswrapper[4799]: I1129 04:51:48.864326 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/41eeb1eb-9b97-41fa-ad22-09d4d4b34abc-metrics-certs\") pod \"openstack-operator-controller-manager-576b8587b8-cftqf\" (UID: \"41eeb1eb-9b97-41fa-ad22-09d4d4b34abc\") " pod="openstack-operators/openstack-operator-controller-manager-576b8587b8-cftqf" Nov 29 04:51:48 crc kubenswrapper[4799]: I1129 04:51:48.864465 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/41eeb1eb-9b97-41fa-ad22-09d4d4b34abc-webhook-certs\") pod \"openstack-operator-controller-manager-576b8587b8-cftqf\" (UID: \"41eeb1eb-9b97-41fa-ad22-09d4d4b34abc\") " pod="openstack-operators/openstack-operator-controller-manager-576b8587b8-cftqf" Nov 29 04:51:48 crc kubenswrapper[4799]: E1129 04:51:48.864622 4799 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 29 04:51:48 crc kubenswrapper[4799]: E1129 04:51:48.864685 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/41eeb1eb-9b97-41fa-ad22-09d4d4b34abc-webhook-certs podName:41eeb1eb-9b97-41fa-ad22-09d4d4b34abc nodeName:}" failed. No retries permitted until 2025-11-29 04:51:56.864665669 +0000 UTC m=+792.507596069 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/41eeb1eb-9b97-41fa-ad22-09d4d4b34abc-webhook-certs") pod "openstack-operator-controller-manager-576b8587b8-cftqf" (UID: "41eeb1eb-9b97-41fa-ad22-09d4d4b34abc") : secret "webhook-server-cert" not found Nov 29 04:51:48 crc kubenswrapper[4799]: E1129 04:51:48.865316 4799 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 29 04:51:48 crc kubenswrapper[4799]: E1129 04:51:48.865520 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/41eeb1eb-9b97-41fa-ad22-09d4d4b34abc-metrics-certs podName:41eeb1eb-9b97-41fa-ad22-09d4d4b34abc nodeName:}" failed. No retries permitted until 2025-11-29 04:51:56.865491338 +0000 UTC m=+792.508421938 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/41eeb1eb-9b97-41fa-ad22-09d4d4b34abc-metrics-certs") pod "openstack-operator-controller-manager-576b8587b8-cftqf" (UID: "41eeb1eb-9b97-41fa-ad22-09d4d4b34abc") : secret "metrics-server-cert" not found Nov 29 04:51:55 crc kubenswrapper[4799]: E1129 04:51:55.727555 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:986861e5a0a9954f63581d9d55a30f8057883cefea489415d76257774526eea3" Nov 29 04:51:55 crc kubenswrapper[4799]: E1129 04:51:55.728633 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:986861e5a0a9954f63581d9d55a30f8057883cefea489415d76257774526eea3,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-j4xpw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-546d4bdf48-8p2z8_openstack-operators(f696db58-bd41-4db0-9ec7-c4a4710e6b8e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 29 04:51:56 crc kubenswrapper[4799]: I1129 04:51:56.224439 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8c9ec062-c2b9-47a6-8d7a-eb9f00e63ef3-cert\") pod \"infra-operator-controller-manager-57548d458d-pn7wx\" (UID: \"8c9ec062-c2b9-47a6-8d7a-eb9f00e63ef3\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-pn7wx" Nov 29 04:51:56 crc kubenswrapper[4799]: I1129 04:51:56.232233 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8c9ec062-c2b9-47a6-8d7a-eb9f00e63ef3-cert\") pod \"infra-operator-controller-manager-57548d458d-pn7wx\" (UID: \"8c9ec062-c2b9-47a6-8d7a-eb9f00e63ef3\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-pn7wx" Nov 29 04:51:56 crc kubenswrapper[4799]: I1129 04:51:56.435383 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-rs8wz" Nov 29 04:51:56 crc kubenswrapper[4799]: I1129 04:51:56.443388 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-pn7wx" Nov 29 04:51:56 crc kubenswrapper[4799]: I1129 04:51:56.528840 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8e0677fe-7f53-4e50-881b-943fffe1c0ff-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4qnwnt\" (UID: \"8e0677fe-7f53-4e50-881b-943fffe1c0ff\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4qnwnt" Nov 29 04:51:56 crc kubenswrapper[4799]: E1129 04:51:56.529301 4799 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 29 04:51:56 crc kubenswrapper[4799]: E1129 04:51:56.529423 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8e0677fe-7f53-4e50-881b-943fffe1c0ff-cert podName:8e0677fe-7f53-4e50-881b-943fffe1c0ff nodeName:}" failed. No retries permitted until 2025-11-29 04:52:12.529391417 +0000 UTC m=+808.172321877 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/8e0677fe-7f53-4e50-881b-943fffe1c0ff-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4qnwnt" (UID: "8e0677fe-7f53-4e50-881b-943fffe1c0ff") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 29 04:51:56 crc kubenswrapper[4799]: I1129 04:51:56.934457 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/41eeb1eb-9b97-41fa-ad22-09d4d4b34abc-webhook-certs\") pod \"openstack-operator-controller-manager-576b8587b8-cftqf\" (UID: \"41eeb1eb-9b97-41fa-ad22-09d4d4b34abc\") " pod="openstack-operators/openstack-operator-controller-manager-576b8587b8-cftqf" Nov 29 04:51:56 crc kubenswrapper[4799]: I1129 04:51:56.934843 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/41eeb1eb-9b97-41fa-ad22-09d4d4b34abc-metrics-certs\") pod \"openstack-operator-controller-manager-576b8587b8-cftqf\" (UID: \"41eeb1eb-9b97-41fa-ad22-09d4d4b34abc\") " pod="openstack-operators/openstack-operator-controller-manager-576b8587b8-cftqf" Nov 29 04:51:56 crc kubenswrapper[4799]: I1129 04:51:56.943633 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/41eeb1eb-9b97-41fa-ad22-09d4d4b34abc-webhook-certs\") pod \"openstack-operator-controller-manager-576b8587b8-cftqf\" (UID: \"41eeb1eb-9b97-41fa-ad22-09d4d4b34abc\") " pod="openstack-operators/openstack-operator-controller-manager-576b8587b8-cftqf" Nov 29 04:51:56 crc kubenswrapper[4799]: I1129 04:51:56.943660 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/41eeb1eb-9b97-41fa-ad22-09d4d4b34abc-metrics-certs\") pod \"openstack-operator-controller-manager-576b8587b8-cftqf\" (UID: \"41eeb1eb-9b97-41fa-ad22-09d4d4b34abc\") " pod="openstack-operators/openstack-operator-controller-manager-576b8587b8-cftqf" Nov 29 04:51:57 crc kubenswrapper[4799]: I1129 04:51:57.130679 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-sfpmg" event={"ID":"7851df67-88d5-45d8-804e-5d3bf3fdb3af","Type":"ContainerStarted","Data":"079ee61c78cf79fc6da6f8545dcd6a5c9bc44ea6e94df1f124f5d80a770de660"} Nov 29 04:51:57 crc kubenswrapper[4799]: I1129 04:51:57.134166 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-wv65n" event={"ID":"06627027-315a-482d-9e26-32c5b4df115a","Type":"ContainerStarted","Data":"f542ac59ad81eebc53e05f7b2514988da63c9877d3777930953ff3e4195d8666"} Nov 29 04:51:57 crc kubenswrapper[4799]: I1129 04:51:57.215253 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-nkvn4" Nov 29 04:51:57 crc kubenswrapper[4799]: I1129 04:51:57.224839 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-576b8587b8-cftqf" Nov 29 04:51:57 crc kubenswrapper[4799]: I1129 04:51:57.836676 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-pn7wx"] Nov 29 04:51:57 crc kubenswrapper[4799]: W1129 04:51:57.901863 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8c9ec062_c2b9_47a6_8d7a_eb9f00e63ef3.slice/crio-199c5c4d22d1ae4d5a925d9c34b85d7822149cabbe6f29a0494f9116f8915c95 WatchSource:0}: Error finding container 199c5c4d22d1ae4d5a925d9c34b85d7822149cabbe6f29a0494f9116f8915c95: Status 404 returned error can't find the container with id 199c5c4d22d1ae4d5a925d9c34b85d7822149cabbe6f29a0494f9116f8915c95 Nov 29 04:51:58 crc kubenswrapper[4799]: I1129 04:51:58.153138 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-lkbfl" event={"ID":"743ba96f-315f-4e0d-abb0-5a6389140c7b","Type":"ContainerStarted","Data":"c81a8dc08724a1387635786acd69a1c9345dc5d85e01807fb7e5881334d8dead"} Nov 29 04:51:58 crc kubenswrapper[4799]: I1129 04:51:58.156763 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-wstbs" event={"ID":"bbe2e2fb-4619-405c-a9ad-4bdf727882cd","Type":"ContainerStarted","Data":"6decb56fc115f3d58ffdaa6962eea7b72372d986d88e9fcde0273a7b19da2213"} Nov 29 04:51:58 crc kubenswrapper[4799]: I1129 04:51:58.157765 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-pn7wx" event={"ID":"8c9ec062-c2b9-47a6-8d7a-eb9f00e63ef3","Type":"ContainerStarted","Data":"199c5c4d22d1ae4d5a925d9c34b85d7822149cabbe6f29a0494f9116f8915c95"} Nov 29 04:51:58 crc kubenswrapper[4799]: I1129 04:51:58.163936 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fsl4h" event={"ID":"225a8e52-78ae-41eb-9c27-64edb57a135e","Type":"ContainerStarted","Data":"199abfe02966ae42c847a074a7fbc8f57c054971beaea709308982edc4811c93"} Nov 29 04:51:58 crc kubenswrapper[4799]: I1129 04:51:58.166923 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-7zddj" event={"ID":"259abc1e-a30c-4a8b-acd1-18052862198e","Type":"ContainerStarted","Data":"8ee08a41e3df22f548a09153ea10c7abe0861c5f1479caad7ed61b09d21fda65"} Nov 29 04:51:58 crc kubenswrapper[4799]: I1129 04:51:58.173650 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-25fzb" event={"ID":"ff72829c-dfd8-4ae2-b508-f89e40f654eb","Type":"ContainerStarted","Data":"6fd63c01843b7f492198988fc6f46524b127618cfa61692b45b01a42bfba3157"} Nov 29 04:51:58 crc kubenswrapper[4799]: I1129 04:51:58.174887 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-qfp99" event={"ID":"888d5a2b-8a13-4042-8c43-59241cfdc088","Type":"ContainerStarted","Data":"81a1cdf833b432bfecd625a099941349afd9e6800131db495ccd45ce3879cffa"} Nov 29 04:51:58 crc kubenswrapper[4799]: I1129 04:51:58.176997 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-84f754f96-t5gjl" event={"ID":"4bf2a041-0f89-480e-88d2-ba630db189c1","Type":"ContainerStarted","Data":"5de42c790857a8e8ddaf7a0cda9f53091b14dbb8e84875eb40b4cf9ab161c3b1"} Nov 29 04:51:58 crc kubenswrapper[4799]: I1129 04:51:58.181232 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-p555g" event={"ID":"2798aeab-f9b9-4728-ad63-186b16afcc31","Type":"ContainerStarted","Data":"bf205560ceca27aa5d995c3eab682d7e9ee6b41f89b3082b65c22c900068c557"} Nov 29 04:51:58 crc kubenswrapper[4799]: I1129 04:51:58.183282 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-6vnhr" event={"ID":"7232eadb-c887-440b-b6f9-bffa4a6dc871","Type":"ContainerStarted","Data":"9c9efbc3bc9a5cb81a71ead4d99e5b4d5f5952fd55301b055455eb614506dc3b"} Nov 29 04:51:58 crc kubenswrapper[4799]: I1129 04:51:58.197499 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fsl4h" podStartSLOduration=4.257737596 podStartE2EDuration="18.197474211s" podCreationTimestamp="2025-11-29 04:51:40 +0000 UTC" firstStartedPulling="2025-11-29 04:51:42.256255077 +0000 UTC m=+777.899185477" lastFinishedPulling="2025-11-29 04:51:56.195991692 +0000 UTC m=+791.838922092" observedRunningTime="2025-11-29 04:51:58.193871957 +0000 UTC m=+793.836802357" watchObservedRunningTime="2025-11-29 04:51:58.197474211 +0000 UTC m=+793.840404611" Nov 29 04:51:58 crc kubenswrapper[4799]: I1129 04:51:58.227342 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-576b8587b8-cftqf"] Nov 29 04:51:58 crc kubenswrapper[4799]: E1129 04:51:58.350610 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"parsing image configuration: Get \\\"https://cdn01.quay.io/quayio-production-s3/sha256/2e/2e965a08b1e129438034a39257c39c4f41bcefa3d8de39716f1e53ad90f967f2?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIATAAF2YHTGR23ZTE6%2F20251129%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20251129T045142Z&X-Amz-Expires=600&X-Amz-SignedHeaders=host&X-Amz-Signature=ba292f783ddde29d42f39baef54928665bba33bb8942cd267e625d402b5f669d®ion=us-east-1&namespace=openstack-k8s-operators&username=openshift-release-dev+ocm_access_1b89217552bc42d1be3fb06a1aed001a&repo_name=neutron-operator&akamai_signature=exp=1764392802~hmac=a85ab262adc9a7f756e25ba39129ce6236b33a2cd435f1795733e94cd99ffae1\\\": remote error: tls: internal error\"" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-pc2zl" podUID="e1d949de-4ce7-4016-83fd-4c0ce003eb0f" Nov 29 04:51:58 crc kubenswrapper[4799]: E1129 04:51:58.612604 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-8p2z8" podUID="f696db58-bd41-4db0-9ec7-c4a4710e6b8e" Nov 29 04:51:59 crc kubenswrapper[4799]: I1129 04:51:59.213132 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-pc2zl" event={"ID":"e1d949de-4ce7-4016-83fd-4c0ce003eb0f","Type":"ContainerStarted","Data":"2ab06f43438c75c4309f294846ec479ea950ffd3caa62eac20ea2cf94a0cc5af"} Nov 29 04:51:59 crc kubenswrapper[4799]: I1129 04:51:59.239183 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-576b8587b8-cftqf" event={"ID":"41eeb1eb-9b97-41fa-ad22-09d4d4b34abc","Type":"ContainerStarted","Data":"2ab889bba39c7c3495ae1c9bedae72f8c044d90360f1342729299503421f3466"} Nov 29 04:51:59 crc kubenswrapper[4799]: I1129 04:51:59.262944 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-sfpmg" event={"ID":"7851df67-88d5-45d8-804e-5d3bf3fdb3af","Type":"ContainerStarted","Data":"0ab93c1be1e6331d83469a24da6a057e4f2a6be7a5ced66181aed5abab9e170a"} Nov 29 04:51:59 crc kubenswrapper[4799]: I1129 04:51:59.263515 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-sfpmg" Nov 29 04:51:59 crc kubenswrapper[4799]: I1129 04:51:59.276444 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-wv65n" event={"ID":"06627027-315a-482d-9e26-32c5b4df115a","Type":"ContainerStarted","Data":"f8a5cea310831e2cf6edfd08865263a0fbc47a0449938b43b96029941e064969"} Nov 29 04:51:59 crc kubenswrapper[4799]: I1129 04:51:59.277304 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-wv65n" Nov 29 04:51:59 crc kubenswrapper[4799]: I1129 04:51:59.291013 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-sfpmg" podStartSLOduration=3.554382778 podStartE2EDuration="19.290995433s" podCreationTimestamp="2025-11-29 04:51:40 +0000 UTC" firstStartedPulling="2025-11-29 04:51:41.984817645 +0000 UTC m=+777.627748035" lastFinishedPulling="2025-11-29 04:51:57.72143029 +0000 UTC m=+793.364360690" observedRunningTime="2025-11-29 04:51:59.289308924 +0000 UTC m=+794.932239324" watchObservedRunningTime="2025-11-29 04:51:59.290995433 +0000 UTC m=+794.933925833" Nov 29 04:51:59 crc kubenswrapper[4799]: I1129 04:51:59.297380 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-wstbs" event={"ID":"bbe2e2fb-4619-405c-a9ad-4bdf727882cd","Type":"ContainerStarted","Data":"5c805f6e70b0973e7fc266004f433f68e3186ceb1497c09b489edbc3a11929e7"} Nov 29 04:51:59 crc kubenswrapper[4799]: I1129 04:51:59.298091 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-wstbs" Nov 29 04:51:59 crc kubenswrapper[4799]: I1129 04:51:59.326921 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-wv65n" podStartSLOduration=3.689430658 podStartE2EDuration="19.326892248s" podCreationTimestamp="2025-11-29 04:51:40 +0000 UTC" firstStartedPulling="2025-11-29 04:51:42.264641172 +0000 UTC m=+777.907571562" lastFinishedPulling="2025-11-29 04:51:57.902102752 +0000 UTC m=+793.545033152" observedRunningTime="2025-11-29 04:51:59.317596042 +0000 UTC m=+794.960526442" watchObservedRunningTime="2025-11-29 04:51:59.326892248 +0000 UTC m=+794.969822648" Nov 29 04:51:59 crc kubenswrapper[4799]: I1129 04:51:59.343226 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-8nl87" event={"ID":"bdcce387-cd82-490e-bd50-cf47158b1ea6","Type":"ContainerStarted","Data":"cdff4a53491f37f67d169e64c123f8a7dc74eb6750ccdaacde6a5cc642ca4abe"} Nov 29 04:51:59 crc kubenswrapper[4799]: I1129 04:51:59.353964 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-8p2z8" event={"ID":"f696db58-bd41-4db0-9ec7-c4a4710e6b8e","Type":"ContainerStarted","Data":"f520c913213d3bdcefcfac1cb88eca785f692eb2cb5d1f0932e39fa537da029d"} Nov 29 04:51:59 crc kubenswrapper[4799]: E1129 04:51:59.355652 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:986861e5a0a9954f63581d9d55a30f8057883cefea489415d76257774526eea3\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-8p2z8" podUID="f696db58-bd41-4db0-9ec7-c4a4710e6b8e" Nov 29 04:51:59 crc kubenswrapper[4799]: I1129 04:51:59.364035 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-wstbs" podStartSLOduration=5.921274574 podStartE2EDuration="19.364017861s" podCreationTimestamp="2025-11-29 04:51:40 +0000 UTC" firstStartedPulling="2025-11-29 04:51:42.249507841 +0000 UTC m=+777.892438251" lastFinishedPulling="2025-11-29 04:51:55.692251138 +0000 UTC m=+791.335181538" observedRunningTime="2025-11-29 04:51:59.359361293 +0000 UTC m=+795.002291693" watchObservedRunningTime="2025-11-29 04:51:59.364017861 +0000 UTC m=+795.006948251" Nov 29 04:51:59 crc kubenswrapper[4799]: I1129 04:51:59.368299 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-6vnhr" event={"ID":"7232eadb-c887-440b-b6f9-bffa4a6dc871","Type":"ContainerStarted","Data":"6b9e8cb0d686b772b65f367aaa4825d43594a8c8c0337447caddf4425c43a8db"} Nov 29 04:51:59 crc kubenswrapper[4799]: I1129 04:51:59.368957 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-6vnhr" Nov 29 04:51:59 crc kubenswrapper[4799]: I1129 04:51:59.377100 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-25fzb" event={"ID":"ff72829c-dfd8-4ae2-b508-f89e40f654eb","Type":"ContainerStarted","Data":"73a4bf585e63f4084c63e27cfca3a12f24cf18b18eaf195e46711217b173314e"} Nov 29 04:51:59 crc kubenswrapper[4799]: I1129 04:51:59.377950 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-25fzb" Nov 29 04:51:59 crc kubenswrapper[4799]: I1129 04:51:59.389907 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kxgw8" event={"ID":"94ec7bc8-8413-4040-90dd-a5c9ce4ca98f","Type":"ContainerStarted","Data":"b0afa40f86f70738d81b8d507d6b851df53dcef37e10455103655dee06bca130"} Nov 29 04:51:59 crc kubenswrapper[4799]: I1129 04:51:59.406481 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-lkbfl" event={"ID":"743ba96f-315f-4e0d-abb0-5a6389140c7b","Type":"ContainerStarted","Data":"c43155a06be9b864154cc9793eb35a7afe8f1203377f32bd525b1806ef88c1a8"} Nov 29 04:51:59 crc kubenswrapper[4799]: I1129 04:51:59.406543 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-lkbfl" Nov 29 04:51:59 crc kubenswrapper[4799]: I1129 04:51:59.424604 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-6vnhr" podStartSLOduration=4.794848957 podStartE2EDuration="19.424575979s" podCreationTimestamp="2025-11-29 04:51:40 +0000 UTC" firstStartedPulling="2025-11-29 04:51:41.062978286 +0000 UTC m=+776.705908686" lastFinishedPulling="2025-11-29 04:51:55.692705308 +0000 UTC m=+791.335635708" observedRunningTime="2025-11-29 04:51:59.415294164 +0000 UTC m=+795.058224564" watchObservedRunningTime="2025-11-29 04:51:59.424575979 +0000 UTC m=+795.067506389" Nov 29 04:51:59 crc kubenswrapper[4799]: I1129 04:51:59.451021 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-25fzb" podStartSLOduration=5.765620684 podStartE2EDuration="19.450986444s" podCreationTimestamp="2025-11-29 04:51:40 +0000 UTC" firstStartedPulling="2025-11-29 04:51:42.006746075 +0000 UTC m=+777.649676475" lastFinishedPulling="2025-11-29 04:51:55.692111835 +0000 UTC m=+791.335042235" observedRunningTime="2025-11-29 04:51:59.439975917 +0000 UTC m=+795.082906327" watchObservedRunningTime="2025-11-29 04:51:59.450986444 +0000 UTC m=+795.093916844" Nov 29 04:52:00 crc kubenswrapper[4799]: I1129 04:52:00.451644 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-p555g" event={"ID":"2798aeab-f9b9-4728-ad63-186b16afcc31","Type":"ContainerStarted","Data":"5e8e3ea85c89811cd63e2e47da8d57f48da5fbef1b56df66e428262ca934d80b"} Nov 29 04:52:00 crc kubenswrapper[4799]: I1129 04:52:00.452652 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-p555g" Nov 29 04:52:00 crc kubenswrapper[4799]: I1129 04:52:00.458454 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-8nl87" event={"ID":"bdcce387-cd82-490e-bd50-cf47158b1ea6","Type":"ContainerStarted","Data":"e623d8bc35a24e604be69cca448cfd6c40533debfb4131c5a439699845361bd9"} Nov 29 04:52:00 crc kubenswrapper[4799]: I1129 04:52:00.459082 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-8nl87" Nov 29 04:52:00 crc kubenswrapper[4799]: I1129 04:52:00.468242 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-7zddj" event={"ID":"259abc1e-a30c-4a8b-acd1-18052862198e","Type":"ContainerStarted","Data":"4be030de49d36f632043ea3492f5488dcaab2e8e2c4cd679dc14c3a85b41bd11"} Nov 29 04:52:00 crc kubenswrapper[4799]: I1129 04:52:00.468772 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-7zddj" Nov 29 04:52:00 crc kubenswrapper[4799]: I1129 04:52:00.469530 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-p555g" podStartSLOduration=6.465252516 podStartE2EDuration="20.469509602s" podCreationTimestamp="2025-11-29 04:51:40 +0000 UTC" firstStartedPulling="2025-11-29 04:51:41.691314919 +0000 UTC m=+777.334245339" lastFinishedPulling="2025-11-29 04:51:55.695572025 +0000 UTC m=+791.338502425" observedRunningTime="2025-11-29 04:52:00.465284234 +0000 UTC m=+796.108214634" watchObservedRunningTime="2025-11-29 04:52:00.469509602 +0000 UTC m=+796.112440002" Nov 29 04:52:00 crc kubenswrapper[4799]: I1129 04:52:00.470702 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-lkbfl" podStartSLOduration=5.993515716 podStartE2EDuration="20.470693739s" podCreationTimestamp="2025-11-29 04:51:40 +0000 UTC" firstStartedPulling="2025-11-29 04:51:41.216238611 +0000 UTC m=+776.859169011" lastFinishedPulling="2025-11-29 04:51:55.693416634 +0000 UTC m=+791.336347034" observedRunningTime="2025-11-29 04:51:59.480658103 +0000 UTC m=+795.123588493" watchObservedRunningTime="2025-11-29 04:52:00.470693739 +0000 UTC m=+796.113624139" Nov 29 04:52:00 crc kubenswrapper[4799]: I1129 04:52:00.473686 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kxgw8" event={"ID":"94ec7bc8-8413-4040-90dd-a5c9ce4ca98f","Type":"ContainerStarted","Data":"096484e24545fb7f6957c57268a4aad4c4410bc888ad9b16e5a61dfd60d23986"} Nov 29 04:52:00 crc kubenswrapper[4799]: I1129 04:52:00.473838 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kxgw8" Nov 29 04:52:00 crc kubenswrapper[4799]: I1129 04:52:00.477225 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-qfp99" event={"ID":"888d5a2b-8a13-4042-8c43-59241cfdc088","Type":"ContainerStarted","Data":"68e52a7988a18e9a847e58ab2b70c35668f5df91171b428d1c1b084cdff7317f"} Nov 29 04:52:00 crc kubenswrapper[4799]: I1129 04:52:00.477383 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-qfp99" Nov 29 04:52:00 crc kubenswrapper[4799]: I1129 04:52:00.482751 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-8nl87" podStartSLOduration=6.501838587 podStartE2EDuration="20.482724589s" podCreationTimestamp="2025-11-29 04:51:40 +0000 UTC" firstStartedPulling="2025-11-29 04:51:42.214519297 +0000 UTC m=+777.857449697" lastFinishedPulling="2025-11-29 04:51:56.195405299 +0000 UTC m=+791.838335699" observedRunningTime="2025-11-29 04:52:00.482552345 +0000 UTC m=+796.125482755" watchObservedRunningTime="2025-11-29 04:52:00.482724589 +0000 UTC m=+796.125654989" Nov 29 04:52:00 crc kubenswrapper[4799]: I1129 04:52:00.483971 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-576b8587b8-cftqf" event={"ID":"41eeb1eb-9b97-41fa-ad22-09d4d4b34abc","Type":"ContainerStarted","Data":"7cf98a6fddd2c42ceb1b26493f5f22e0e313b7d0be53bdef4cab4b13c84eb469"} Nov 29 04:52:00 crc kubenswrapper[4799]: I1129 04:52:00.484057 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-576b8587b8-cftqf" Nov 29 04:52:00 crc kubenswrapper[4799]: I1129 04:52:00.487257 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-84f754f96-t5gjl" event={"ID":"4bf2a041-0f89-480e-88d2-ba630db189c1","Type":"ContainerStarted","Data":"3c02a4e669cc1172055e8b9d74196129627f991c37db907ae4e7a61929284fc8"} Nov 29 04:52:00 crc kubenswrapper[4799]: E1129 04:52:00.491529 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:986861e5a0a9954f63581d9d55a30f8057883cefea489415d76257774526eea3\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-8p2z8" podUID="f696db58-bd41-4db0-9ec7-c4a4710e6b8e" Nov 29 04:52:00 crc kubenswrapper[4799]: I1129 04:52:00.511996 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kxgw8" podStartSLOduration=5.9551278530000005 podStartE2EDuration="20.511971549s" podCreationTimestamp="2025-11-29 04:51:40 +0000 UTC" firstStartedPulling="2025-11-29 04:51:41.639179947 +0000 UTC m=+777.282110347" lastFinishedPulling="2025-11-29 04:51:56.196023633 +0000 UTC m=+791.838954043" observedRunningTime="2025-11-29 04:52:00.507745211 +0000 UTC m=+796.150675631" watchObservedRunningTime="2025-11-29 04:52:00.511971549 +0000 UTC m=+796.154901949" Nov 29 04:52:00 crc kubenswrapper[4799]: I1129 04:52:00.539158 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-7zddj" podStartSLOduration=6.859889664 podStartE2EDuration="20.53913231s" podCreationTimestamp="2025-11-29 04:51:40 +0000 UTC" firstStartedPulling="2025-11-29 04:51:42.014176238 +0000 UTC m=+777.657106638" lastFinishedPulling="2025-11-29 04:51:55.693418884 +0000 UTC m=+791.336349284" observedRunningTime="2025-11-29 04:52:00.530986242 +0000 UTC m=+796.173916662" watchObservedRunningTime="2025-11-29 04:52:00.53913231 +0000 UTC m=+796.182062710" Nov 29 04:52:00 crc kubenswrapper[4799]: I1129 04:52:00.555356 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-qfp99" podStartSLOduration=6.9095902890000005 podStartE2EDuration="20.555331217s" podCreationTimestamp="2025-11-29 04:51:40 +0000 UTC" firstStartedPulling="2025-11-29 04:51:42.23958597 +0000 UTC m=+777.882516360" lastFinishedPulling="2025-11-29 04:51:55.885326888 +0000 UTC m=+791.528257288" observedRunningTime="2025-11-29 04:52:00.553280329 +0000 UTC m=+796.196210729" watchObservedRunningTime="2025-11-29 04:52:00.555331217 +0000 UTC m=+796.198261617" Nov 29 04:52:00 crc kubenswrapper[4799]: I1129 04:52:00.566923 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-84f754f96-t5gjl" Nov 29 04:52:00 crc kubenswrapper[4799]: I1129 04:52:00.623933 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-84f754f96-t5gjl" podStartSLOduration=6.595389452 podStartE2EDuration="20.623906962s" podCreationTimestamp="2025-11-29 04:51:40 +0000 UTC" firstStartedPulling="2025-11-29 04:51:41.665397076 +0000 UTC m=+777.308327476" lastFinishedPulling="2025-11-29 04:51:55.693914586 +0000 UTC m=+791.336844986" observedRunningTime="2025-11-29 04:52:00.622379017 +0000 UTC m=+796.265309417" watchObservedRunningTime="2025-11-29 04:52:00.623906962 +0000 UTC m=+796.266837362" Nov 29 04:52:00 crc kubenswrapper[4799]: I1129 04:52:00.680396 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-576b8587b8-cftqf" podStartSLOduration=20.680371415 podStartE2EDuration="20.680371415s" podCreationTimestamp="2025-11-29 04:51:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:52:00.672324978 +0000 UTC m=+796.315255378" watchObservedRunningTime="2025-11-29 04:52:00.680371415 +0000 UTC m=+796.323301815" Nov 29 04:52:01 crc kubenswrapper[4799]: I1129 04:52:01.079330 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-wv65n" Nov 29 04:52:07 crc kubenswrapper[4799]: I1129 04:52:07.232783 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-576b8587b8-cftqf" Nov 29 04:52:08 crc kubenswrapper[4799]: I1129 04:52:08.569752 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-pn7wx" event={"ID":"8c9ec062-c2b9-47a6-8d7a-eb9f00e63ef3","Type":"ContainerStarted","Data":"8c8f74d4c52de6b4f33633ace0111420c2b60da330ca9a14d8750a19529cd19a"} Nov 29 04:52:08 crc kubenswrapper[4799]: I1129 04:52:08.572351 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-tbth4" event={"ID":"1bb0507a-76c1-4a53-971c-dce960b71242","Type":"ContainerStarted","Data":"d65107bdb125dbe35bcd2ec13a662719e9b34eaf57a69fb56328f30be822ec4c"} Nov 29 04:52:08 crc kubenswrapper[4799]: I1129 04:52:08.572379 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-tbth4" event={"ID":"1bb0507a-76c1-4a53-971c-dce960b71242","Type":"ContainerStarted","Data":"33893f17d6ce8d0de5432b9af0d240baf497a548e2ec98dbcabc58c2a105cc4e"} Nov 29 04:52:08 crc kubenswrapper[4799]: I1129 04:52:08.573641 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-tbth4" Nov 29 04:52:08 crc kubenswrapper[4799]: I1129 04:52:08.575725 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-pc2zl" event={"ID":"e1d949de-4ce7-4016-83fd-4c0ce003eb0f","Type":"ContainerStarted","Data":"2bfd9dd7a7d064699ec6122e3a7c75c668d2a6d6b2ea5cb73135fd3bfebd537c"} Nov 29 04:52:08 crc kubenswrapper[4799]: I1129 04:52:08.576154 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-pc2zl" Nov 29 04:52:08 crc kubenswrapper[4799]: I1129 04:52:08.579248 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-v4kv7" event={"ID":"1edee957-6d50-48c2-bccc-e6702bcb5ce0","Type":"ContainerStarted","Data":"dd1116f37ae8163bb0b413a00a864a0a8746d1c214dd86322e34aa3e43c323f1"} Nov 29 04:52:08 crc kubenswrapper[4799]: I1129 04:52:08.579310 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-v4kv7" event={"ID":"1edee957-6d50-48c2-bccc-e6702bcb5ce0","Type":"ContainerStarted","Data":"ac72d4c696f19cca267efcca40401faf37f1eebd2454884bc39fd09e2ac7cd9e"} Nov 29 04:52:08 crc kubenswrapper[4799]: I1129 04:52:08.579567 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-v4kv7" Nov 29 04:52:08 crc kubenswrapper[4799]: I1129 04:52:08.582208 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-4xvn4" event={"ID":"ac90f9ba-d2ef-44a1-84e7-8e809b5eef13","Type":"ContainerStarted","Data":"fc0cc2e52fd18662cadb260b857d91a09f9089677183f36032b45c6a3d2ff8f1"} Nov 29 04:52:08 crc kubenswrapper[4799]: I1129 04:52:08.583507 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-pkrhh" event={"ID":"89038368-29a3-417e-8ffa-819e5901a52a","Type":"ContainerStarted","Data":"fc98fe7c17ce3be4eacc1a49686d656d1819c9e14da267bcf13778ae88569b29"} Nov 29 04:52:08 crc kubenswrapper[4799]: I1129 04:52:08.584737 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-4p69t" event={"ID":"861aea23-001f-4259-b69a-64b95db49c73","Type":"ContainerStarted","Data":"d6e40e10b94486c927472d6d62972e4a14a73a109ada30c411755b800e3ac289"} Nov 29 04:52:08 crc kubenswrapper[4799]: I1129 04:52:08.596854 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-tbth4" podStartSLOduration=2.8353469650000003 podStartE2EDuration="28.596833077s" podCreationTimestamp="2025-11-29 04:51:40 +0000 UTC" firstStartedPulling="2025-11-29 04:51:42.270638812 +0000 UTC m=+777.913569212" lastFinishedPulling="2025-11-29 04:52:08.032124924 +0000 UTC m=+803.675055324" observedRunningTime="2025-11-29 04:52:08.591574716 +0000 UTC m=+804.234505116" watchObservedRunningTime="2025-11-29 04:52:08.596833077 +0000 UTC m=+804.239763477" Nov 29 04:52:08 crc kubenswrapper[4799]: I1129 04:52:08.647480 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-pc2zl" podStartSLOduration=2.616014065 podStartE2EDuration="28.647458135s" podCreationTimestamp="2025-11-29 04:51:40 +0000 UTC" firstStartedPulling="2025-11-29 04:51:42.004750129 +0000 UTC m=+777.647680529" lastFinishedPulling="2025-11-29 04:52:08.036194199 +0000 UTC m=+803.679124599" observedRunningTime="2025-11-29 04:52:08.639330836 +0000 UTC m=+804.282261236" watchObservedRunningTime="2025-11-29 04:52:08.647458135 +0000 UTC m=+804.290388525" Nov 29 04:52:08 crc kubenswrapper[4799]: I1129 04:52:08.660342 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-v4kv7" podStartSLOduration=2.926633989 podStartE2EDuration="28.660313404s" podCreationTimestamp="2025-11-29 04:51:40 +0000 UTC" firstStartedPulling="2025-11-29 04:51:42.282279793 +0000 UTC m=+777.925210183" lastFinishedPulling="2025-11-29 04:52:08.015959188 +0000 UTC m=+803.658889598" observedRunningTime="2025-11-29 04:52:08.656963346 +0000 UTC m=+804.299893746" watchObservedRunningTime="2025-11-29 04:52:08.660313404 +0000 UTC m=+804.303243804" Nov 29 04:52:09 crc kubenswrapper[4799]: I1129 04:52:09.593687 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-4xvn4" event={"ID":"ac90f9ba-d2ef-44a1-84e7-8e809b5eef13","Type":"ContainerStarted","Data":"f4a30cb25ee48d273db7532f637dcae7e4babec8b04e970b26f574e25c49256d"} Nov 29 04:52:09 crc kubenswrapper[4799]: I1129 04:52:09.594061 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-4xvn4" Nov 29 04:52:09 crc kubenswrapper[4799]: I1129 04:52:09.596634 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-pkrhh" event={"ID":"89038368-29a3-417e-8ffa-819e5901a52a","Type":"ContainerStarted","Data":"85e27219bd2ebaf8dce6b64b7bbb324877c2be91c85d9138d37ef060638b6b8a"} Nov 29 04:52:09 crc kubenswrapper[4799]: I1129 04:52:09.596779 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-pkrhh" Nov 29 04:52:09 crc kubenswrapper[4799]: I1129 04:52:09.598941 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-4p69t" event={"ID":"861aea23-001f-4259-b69a-64b95db49c73","Type":"ContainerStarted","Data":"ad757e7176841f928cb74d4c2b024668fe7c9ba053e3f36b9ed5e45c09d597ab"} Nov 29 04:52:09 crc kubenswrapper[4799]: I1129 04:52:09.599068 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-4p69t" Nov 29 04:52:09 crc kubenswrapper[4799]: I1129 04:52:09.604993 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-pn7wx" event={"ID":"8c9ec062-c2b9-47a6-8d7a-eb9f00e63ef3","Type":"ContainerStarted","Data":"87f3f7ae44929af690fb94ba2e1ee6e794fca47a84204da12c57eefcb33405dc"} Nov 29 04:52:09 crc kubenswrapper[4799]: I1129 04:52:09.605810 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-pn7wx" Nov 29 04:52:09 crc kubenswrapper[4799]: I1129 04:52:09.625154 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-4xvn4" podStartSLOduration=3.878859416 podStartE2EDuration="29.625110834s" podCreationTimestamp="2025-11-29 04:51:40 +0000 UTC" firstStartedPulling="2025-11-29 04:51:42.281528755 +0000 UTC m=+777.924459155" lastFinishedPulling="2025-11-29 04:52:08.027780173 +0000 UTC m=+803.670710573" observedRunningTime="2025-11-29 04:52:09.62176458 +0000 UTC m=+805.264694980" watchObservedRunningTime="2025-11-29 04:52:09.625110834 +0000 UTC m=+805.268041234" Nov 29 04:52:09 crc kubenswrapper[4799]: I1129 04:52:09.652885 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-pkrhh" podStartSLOduration=3.855454081 podStartE2EDuration="29.652854699s" podCreationTimestamp="2025-11-29 04:51:40 +0000 UTC" firstStartedPulling="2025-11-29 04:51:42.264691523 +0000 UTC m=+777.907621923" lastFinishedPulling="2025-11-29 04:52:08.062092141 +0000 UTC m=+803.705022541" observedRunningTime="2025-11-29 04:52:09.646231786 +0000 UTC m=+805.289162206" watchObservedRunningTime="2025-11-29 04:52:09.652854699 +0000 UTC m=+805.295785099" Nov 29 04:52:09 crc kubenswrapper[4799]: I1129 04:52:09.721183 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-57548d458d-pn7wx" podStartSLOduration=19.624357759 podStartE2EDuration="29.721160419s" podCreationTimestamp="2025-11-29 04:51:40 +0000 UTC" firstStartedPulling="2025-11-29 04:51:57.93427025 +0000 UTC m=+793.577200650" lastFinishedPulling="2025-11-29 04:52:08.03107291 +0000 UTC m=+803.674003310" observedRunningTime="2025-11-29 04:52:09.720676688 +0000 UTC m=+805.363607078" watchObservedRunningTime="2025-11-29 04:52:09.721160419 +0000 UTC m=+805.364090819" Nov 29 04:52:09 crc kubenswrapper[4799]: I1129 04:52:09.724844 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-4p69t" podStartSLOduration=3.958125327 podStartE2EDuration="29.72483538s" podCreationTimestamp="2025-11-29 04:51:40 +0000 UTC" firstStartedPulling="2025-11-29 04:51:42.264804357 +0000 UTC m=+777.907734757" lastFinishedPulling="2025-11-29 04:52:08.03151439 +0000 UTC m=+803.674444810" observedRunningTime="2025-11-29 04:52:09.692087991 +0000 UTC m=+805.335018391" watchObservedRunningTime="2025-11-29 04:52:09.72483538 +0000 UTC m=+805.367765780" Nov 29 04:52:10 crc kubenswrapper[4799]: I1129 04:52:10.377134 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-6vnhr" Nov 29 04:52:10 crc kubenswrapper[4799]: I1129 04:52:10.460158 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-lkbfl" Nov 29 04:52:10 crc kubenswrapper[4799]: I1129 04:52:10.483781 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-p555g" Nov 29 04:52:10 crc kubenswrapper[4799]: I1129 04:52:10.523538 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kxgw8" Nov 29 04:52:10 crc kubenswrapper[4799]: I1129 04:52:10.552065 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-25fzb" Nov 29 04:52:10 crc kubenswrapper[4799]: I1129 04:52:10.571600 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-84f754f96-t5gjl" Nov 29 04:52:10 crc kubenswrapper[4799]: I1129 04:52:10.605469 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-sfpmg" Nov 29 04:52:10 crc kubenswrapper[4799]: I1129 04:52:10.683771 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-7zddj" Nov 29 04:52:10 crc kubenswrapper[4799]: I1129 04:52:10.799007 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-8nl87" Nov 29 04:52:10 crc kubenswrapper[4799]: I1129 04:52:10.871940 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-wstbs" Nov 29 04:52:11 crc kubenswrapper[4799]: I1129 04:52:11.236369 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-qfp99" Nov 29 04:52:12 crc kubenswrapper[4799]: I1129 04:52:12.541418 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8e0677fe-7f53-4e50-881b-943fffe1c0ff-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4qnwnt\" (UID: \"8e0677fe-7f53-4e50-881b-943fffe1c0ff\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4qnwnt" Nov 29 04:52:12 crc kubenswrapper[4799]: I1129 04:52:12.554674 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8e0677fe-7f53-4e50-881b-943fffe1c0ff-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4qnwnt\" (UID: \"8e0677fe-7f53-4e50-881b-943fffe1c0ff\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4qnwnt" Nov 29 04:52:12 crc kubenswrapper[4799]: I1129 04:52:12.680229 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-lpq6g" Nov 29 04:52:12 crc kubenswrapper[4799]: I1129 04:52:12.689118 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4qnwnt" Nov 29 04:52:13 crc kubenswrapper[4799]: I1129 04:52:13.100378 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4qnwnt"] Nov 29 04:52:13 crc kubenswrapper[4799]: W1129 04:52:13.110294 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8e0677fe_7f53_4e50_881b_943fffe1c0ff.slice/crio-fedddaf4843e75a8e03d811501b670af781d3da21d8ddf28732e6222cd406ecf WatchSource:0}: Error finding container fedddaf4843e75a8e03d811501b670af781d3da21d8ddf28732e6222cd406ecf: Status 404 returned error can't find the container with id fedddaf4843e75a8e03d811501b670af781d3da21d8ddf28732e6222cd406ecf Nov 29 04:52:13 crc kubenswrapper[4799]: I1129 04:52:13.635577 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4qnwnt" event={"ID":"8e0677fe-7f53-4e50-881b-943fffe1c0ff","Type":"ContainerStarted","Data":"fedddaf4843e75a8e03d811501b670af781d3da21d8ddf28732e6222cd406ecf"} Nov 29 04:52:16 crc kubenswrapper[4799]: I1129 04:52:16.454444 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-57548d458d-pn7wx" Nov 29 04:52:20 crc kubenswrapper[4799]: I1129 04:52:20.656191 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-pc2zl" Nov 29 04:52:20 crc kubenswrapper[4799]: I1129 04:52:20.707128 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-pkrhh" Nov 29 04:52:20 crc kubenswrapper[4799]: I1129 04:52:20.959047 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-4p69t" Nov 29 04:52:21 crc kubenswrapper[4799]: I1129 04:52:21.009183 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-4xvn4" Nov 29 04:52:21 crc kubenswrapper[4799]: I1129 04:52:21.112523 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-tbth4" Nov 29 04:52:21 crc kubenswrapper[4799]: I1129 04:52:21.221310 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-v4kv7" Nov 29 04:52:25 crc kubenswrapper[4799]: E1129 04:52:25.208172 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = writing blob: storing blob to file \"/var/tmp/container_images_storage2543173169/1\": happened during read: context canceled" image="quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:14cfad6ea2e7f7ecc4cb2aafceb9c61514b3d04b66668832d1e4ac3b19f1ab81" Nov 29 04:52:25 crc kubenswrapper[4799]: E1129 04:52:25.208962 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:14cfad6ea2e7f7ecc4cb2aafceb9c61514b3d04b66668832d1e4ac3b19f1ab81,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter:v0.15.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_API_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_PROC_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-processor:current,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler:release-0.7.12,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter:v1.5.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter:v1.10.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine:current-podified,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-w54bx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-64bc77cfd4qnwnt_openstack-operators(8e0677fe-7f53-4e50-881b-943fffe1c0ff): ErrImagePull: rpc error: code = Canceled desc = writing blob: storing blob to file \"/var/tmp/container_images_storage2543173169/1\": happened during read: context canceled" logger="UnhandledError" Nov 29 04:52:26 crc kubenswrapper[4799]: E1129 04:52:26.196491 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = writing blob: storing blob to file \\\"/var/tmp/container_images_storage2543173169/1\\\": happened during read: context canceled\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4qnwnt" podUID="8e0677fe-7f53-4e50-881b-943fffe1c0ff" Nov 29 04:52:26 crc kubenswrapper[4799]: I1129 04:52:26.768113 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4qnwnt" event={"ID":"8e0677fe-7f53-4e50-881b-943fffe1c0ff","Type":"ContainerStarted","Data":"2a1545b7f7d6d2716bea6b9223b1369889cb4d82d2ee20614c7a749d0b08c68a"} Nov 29 04:52:26 crc kubenswrapper[4799]: I1129 04:52:26.769861 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-8p2z8" event={"ID":"f696db58-bd41-4db0-9ec7-c4a4710e6b8e","Type":"ContainerStarted","Data":"0615216db73cab2e2e191bdcd84161223074df2302466a8337a0b03a616fe3ac"} Nov 29 04:52:26 crc kubenswrapper[4799]: I1129 04:52:26.770134 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-8p2z8" Nov 29 04:52:26 crc kubenswrapper[4799]: E1129 04:52:26.770348 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:14cfad6ea2e7f7ecc4cb2aafceb9c61514b3d04b66668832d1e4ac3b19f1ab81\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4qnwnt" podUID="8e0677fe-7f53-4e50-881b-943fffe1c0ff" Nov 29 04:52:26 crc kubenswrapper[4799]: I1129 04:52:26.819396 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-8p2z8" podStartSLOduration=2.9473906420000002 podStartE2EDuration="46.819368415s" podCreationTimestamp="2025-11-29 04:51:40 +0000 UTC" firstStartedPulling="2025-11-29 04:51:41.691204026 +0000 UTC m=+777.334134426" lastFinishedPulling="2025-11-29 04:52:25.563181799 +0000 UTC m=+821.206112199" observedRunningTime="2025-11-29 04:52:26.816201127 +0000 UTC m=+822.459131527" watchObservedRunningTime="2025-11-29 04:52:26.819368415 +0000 UTC m=+822.462298805" Nov 29 04:52:27 crc kubenswrapper[4799]: E1129 04:52:27.778579 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:14cfad6ea2e7f7ecc4cb2aafceb9c61514b3d04b66668832d1e4ac3b19f1ab81\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4qnwnt" podUID="8e0677fe-7f53-4e50-881b-943fffe1c0ff" Nov 29 04:52:40 crc kubenswrapper[4799]: I1129 04:52:40.597212 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-8p2z8" Nov 29 04:52:43 crc kubenswrapper[4799]: I1129 04:52:43.923831 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4qnwnt" event={"ID":"8e0677fe-7f53-4e50-881b-943fffe1c0ff","Type":"ContainerStarted","Data":"23279abb2bde115011422cb67256d49fbfe80ba47ddadf8721cd594bef96b302"} Nov 29 04:52:43 crc kubenswrapper[4799]: I1129 04:52:43.924356 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4qnwnt" Nov 29 04:52:43 crc kubenswrapper[4799]: I1129 04:52:43.952881 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4qnwnt" podStartSLOduration=33.86137305 podStartE2EDuration="1m3.952852248s" podCreationTimestamp="2025-11-29 04:51:40 +0000 UTC" firstStartedPulling="2025-11-29 04:52:13.112137796 +0000 UTC m=+808.755068196" lastFinishedPulling="2025-11-29 04:52:43.203617004 +0000 UTC m=+838.846547394" observedRunningTime="2025-11-29 04:52:43.950929761 +0000 UTC m=+839.593860151" watchObservedRunningTime="2025-11-29 04:52:43.952852248 +0000 UTC m=+839.595782648" Nov 29 04:52:52 crc kubenswrapper[4799]: I1129 04:52:52.698336 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4qnwnt" Nov 29 04:53:07 crc kubenswrapper[4799]: I1129 04:53:07.551549 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-psbc4"] Nov 29 04:53:07 crc kubenswrapper[4799]: I1129 04:53:07.557373 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-psbc4" Nov 29 04:53:07 crc kubenswrapper[4799]: I1129 04:53:07.564781 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-psbc4"] Nov 29 04:53:07 crc kubenswrapper[4799]: I1129 04:53:07.565347 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Nov 29 04:53:07 crc kubenswrapper[4799]: I1129 04:53:07.565564 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Nov 29 04:53:07 crc kubenswrapper[4799]: I1129 04:53:07.565773 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-dnwxz" Nov 29 04:53:07 crc kubenswrapper[4799]: I1129 04:53:07.565926 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Nov 29 04:53:07 crc kubenswrapper[4799]: I1129 04:53:07.646891 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99c095ca-db71-477a-9c51-3df3d2838144-config\") pod \"dnsmasq-dns-675f4bcbfc-psbc4\" (UID: \"99c095ca-db71-477a-9c51-3df3d2838144\") " pod="openstack/dnsmasq-dns-675f4bcbfc-psbc4" Nov 29 04:53:07 crc kubenswrapper[4799]: I1129 04:53:07.647426 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4g6rd\" (UniqueName: \"kubernetes.io/projected/99c095ca-db71-477a-9c51-3df3d2838144-kube-api-access-4g6rd\") pod \"dnsmasq-dns-675f4bcbfc-psbc4\" (UID: \"99c095ca-db71-477a-9c51-3df3d2838144\") " pod="openstack/dnsmasq-dns-675f4bcbfc-psbc4" Nov 29 04:53:07 crc kubenswrapper[4799]: I1129 04:53:07.646934 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-vv6cv"] Nov 29 04:53:07 crc kubenswrapper[4799]: I1129 04:53:07.649203 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-vv6cv" Nov 29 04:53:07 crc kubenswrapper[4799]: I1129 04:53:07.658620 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Nov 29 04:53:07 crc kubenswrapper[4799]: I1129 04:53:07.684153 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-vv6cv"] Nov 29 04:53:07 crc kubenswrapper[4799]: I1129 04:53:07.749254 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99c095ca-db71-477a-9c51-3df3d2838144-config\") pod \"dnsmasq-dns-675f4bcbfc-psbc4\" (UID: \"99c095ca-db71-477a-9c51-3df3d2838144\") " pod="openstack/dnsmasq-dns-675f4bcbfc-psbc4" Nov 29 04:53:07 crc kubenswrapper[4799]: I1129 04:53:07.749315 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4g6rd\" (UniqueName: \"kubernetes.io/projected/99c095ca-db71-477a-9c51-3df3d2838144-kube-api-access-4g6rd\") pod \"dnsmasq-dns-675f4bcbfc-psbc4\" (UID: \"99c095ca-db71-477a-9c51-3df3d2838144\") " pod="openstack/dnsmasq-dns-675f4bcbfc-psbc4" Nov 29 04:53:07 crc kubenswrapper[4799]: I1129 04:53:07.750881 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99c095ca-db71-477a-9c51-3df3d2838144-config\") pod \"dnsmasq-dns-675f4bcbfc-psbc4\" (UID: \"99c095ca-db71-477a-9c51-3df3d2838144\") " pod="openstack/dnsmasq-dns-675f4bcbfc-psbc4" Nov 29 04:53:07 crc kubenswrapper[4799]: I1129 04:53:07.772051 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4g6rd\" (UniqueName: \"kubernetes.io/projected/99c095ca-db71-477a-9c51-3df3d2838144-kube-api-access-4g6rd\") pod \"dnsmasq-dns-675f4bcbfc-psbc4\" (UID: \"99c095ca-db71-477a-9c51-3df3d2838144\") " pod="openstack/dnsmasq-dns-675f4bcbfc-psbc4" Nov 29 04:53:07 crc kubenswrapper[4799]: I1129 04:53:07.850484 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pjhh\" (UniqueName: \"kubernetes.io/projected/a0f5fc9f-1ab2-42dc-9061-da1ae18610f1-kube-api-access-4pjhh\") pod \"dnsmasq-dns-78dd6ddcc-vv6cv\" (UID: \"a0f5fc9f-1ab2-42dc-9061-da1ae18610f1\") " pod="openstack/dnsmasq-dns-78dd6ddcc-vv6cv" Nov 29 04:53:07 crc kubenswrapper[4799]: I1129 04:53:07.850559 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0f5fc9f-1ab2-42dc-9061-da1ae18610f1-config\") pod \"dnsmasq-dns-78dd6ddcc-vv6cv\" (UID: \"a0f5fc9f-1ab2-42dc-9061-da1ae18610f1\") " pod="openstack/dnsmasq-dns-78dd6ddcc-vv6cv" Nov 29 04:53:07 crc kubenswrapper[4799]: I1129 04:53:07.850598 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a0f5fc9f-1ab2-42dc-9061-da1ae18610f1-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-vv6cv\" (UID: \"a0f5fc9f-1ab2-42dc-9061-da1ae18610f1\") " pod="openstack/dnsmasq-dns-78dd6ddcc-vv6cv" Nov 29 04:53:07 crc kubenswrapper[4799]: I1129 04:53:07.900678 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-psbc4" Nov 29 04:53:07 crc kubenswrapper[4799]: I1129 04:53:07.952473 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pjhh\" (UniqueName: \"kubernetes.io/projected/a0f5fc9f-1ab2-42dc-9061-da1ae18610f1-kube-api-access-4pjhh\") pod \"dnsmasq-dns-78dd6ddcc-vv6cv\" (UID: \"a0f5fc9f-1ab2-42dc-9061-da1ae18610f1\") " pod="openstack/dnsmasq-dns-78dd6ddcc-vv6cv" Nov 29 04:53:07 crc kubenswrapper[4799]: I1129 04:53:07.952934 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0f5fc9f-1ab2-42dc-9061-da1ae18610f1-config\") pod \"dnsmasq-dns-78dd6ddcc-vv6cv\" (UID: \"a0f5fc9f-1ab2-42dc-9061-da1ae18610f1\") " pod="openstack/dnsmasq-dns-78dd6ddcc-vv6cv" Nov 29 04:53:07 crc kubenswrapper[4799]: I1129 04:53:07.952971 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a0f5fc9f-1ab2-42dc-9061-da1ae18610f1-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-vv6cv\" (UID: \"a0f5fc9f-1ab2-42dc-9061-da1ae18610f1\") " pod="openstack/dnsmasq-dns-78dd6ddcc-vv6cv" Nov 29 04:53:07 crc kubenswrapper[4799]: I1129 04:53:07.954280 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a0f5fc9f-1ab2-42dc-9061-da1ae18610f1-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-vv6cv\" (UID: \"a0f5fc9f-1ab2-42dc-9061-da1ae18610f1\") " pod="openstack/dnsmasq-dns-78dd6ddcc-vv6cv" Nov 29 04:53:07 crc kubenswrapper[4799]: I1129 04:53:07.954310 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0f5fc9f-1ab2-42dc-9061-da1ae18610f1-config\") pod \"dnsmasq-dns-78dd6ddcc-vv6cv\" (UID: \"a0f5fc9f-1ab2-42dc-9061-da1ae18610f1\") " pod="openstack/dnsmasq-dns-78dd6ddcc-vv6cv" Nov 29 04:53:07 crc kubenswrapper[4799]: I1129 04:53:07.974078 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pjhh\" (UniqueName: \"kubernetes.io/projected/a0f5fc9f-1ab2-42dc-9061-da1ae18610f1-kube-api-access-4pjhh\") pod \"dnsmasq-dns-78dd6ddcc-vv6cv\" (UID: \"a0f5fc9f-1ab2-42dc-9061-da1ae18610f1\") " pod="openstack/dnsmasq-dns-78dd6ddcc-vv6cv" Nov 29 04:53:08 crc kubenswrapper[4799]: I1129 04:53:08.009393 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-vv6cv" Nov 29 04:53:08 crc kubenswrapper[4799]: I1129 04:53:08.327949 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-psbc4"] Nov 29 04:53:08 crc kubenswrapper[4799]: I1129 04:53:08.335235 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 29 04:53:08 crc kubenswrapper[4799]: I1129 04:53:08.443351 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-vv6cv"] Nov 29 04:53:08 crc kubenswrapper[4799]: W1129 04:53:08.444978 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda0f5fc9f_1ab2_42dc_9061_da1ae18610f1.slice/crio-486a647bff80499971d4e626bb10a5d919d6a4f6541a0060539c166ec316d912 WatchSource:0}: Error finding container 486a647bff80499971d4e626bb10a5d919d6a4f6541a0060539c166ec316d912: Status 404 returned error can't find the container with id 486a647bff80499971d4e626bb10a5d919d6a4f6541a0060539c166ec316d912 Nov 29 04:53:09 crc kubenswrapper[4799]: I1129 04:53:09.153536 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-vv6cv" event={"ID":"a0f5fc9f-1ab2-42dc-9061-da1ae18610f1","Type":"ContainerStarted","Data":"486a647bff80499971d4e626bb10a5d919d6a4f6541a0060539c166ec316d912"} Nov 29 04:53:09 crc kubenswrapper[4799]: I1129 04:53:09.154403 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-psbc4" event={"ID":"99c095ca-db71-477a-9c51-3df3d2838144","Type":"ContainerStarted","Data":"2b37b1b7b391fb353165defa9aefe2718ce43a42c7382dfbc9858a3f47a69028"} Nov 29 04:53:10 crc kubenswrapper[4799]: I1129 04:53:10.252836 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-psbc4"] Nov 29 04:53:10 crc kubenswrapper[4799]: I1129 04:53:10.299560 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-9q9hd"] Nov 29 04:53:10 crc kubenswrapper[4799]: I1129 04:53:10.300776 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-9q9hd" Nov 29 04:53:10 crc kubenswrapper[4799]: I1129 04:53:10.307781 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-9q9hd"] Nov 29 04:53:10 crc kubenswrapper[4799]: I1129 04:53:10.503912 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccqck\" (UniqueName: \"kubernetes.io/projected/da2af6d0-1824-4820-84eb-0c491cb777a9-kube-api-access-ccqck\") pod \"dnsmasq-dns-666b6646f7-9q9hd\" (UID: \"da2af6d0-1824-4820-84eb-0c491cb777a9\") " pod="openstack/dnsmasq-dns-666b6646f7-9q9hd" Nov 29 04:53:10 crc kubenswrapper[4799]: I1129 04:53:10.504539 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da2af6d0-1824-4820-84eb-0c491cb777a9-dns-svc\") pod \"dnsmasq-dns-666b6646f7-9q9hd\" (UID: \"da2af6d0-1824-4820-84eb-0c491cb777a9\") " pod="openstack/dnsmasq-dns-666b6646f7-9q9hd" Nov 29 04:53:10 crc kubenswrapper[4799]: I1129 04:53:10.504574 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da2af6d0-1824-4820-84eb-0c491cb777a9-config\") pod \"dnsmasq-dns-666b6646f7-9q9hd\" (UID: \"da2af6d0-1824-4820-84eb-0c491cb777a9\") " pod="openstack/dnsmasq-dns-666b6646f7-9q9hd" Nov 29 04:53:10 crc kubenswrapper[4799]: I1129 04:53:10.584042 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-vv6cv"] Nov 29 04:53:10 crc kubenswrapper[4799]: I1129 04:53:10.606305 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccqck\" (UniqueName: \"kubernetes.io/projected/da2af6d0-1824-4820-84eb-0c491cb777a9-kube-api-access-ccqck\") pod \"dnsmasq-dns-666b6646f7-9q9hd\" (UID: \"da2af6d0-1824-4820-84eb-0c491cb777a9\") " pod="openstack/dnsmasq-dns-666b6646f7-9q9hd" Nov 29 04:53:10 crc kubenswrapper[4799]: I1129 04:53:10.606428 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da2af6d0-1824-4820-84eb-0c491cb777a9-dns-svc\") pod \"dnsmasq-dns-666b6646f7-9q9hd\" (UID: \"da2af6d0-1824-4820-84eb-0c491cb777a9\") " pod="openstack/dnsmasq-dns-666b6646f7-9q9hd" Nov 29 04:53:10 crc kubenswrapper[4799]: I1129 04:53:10.606468 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da2af6d0-1824-4820-84eb-0c491cb777a9-config\") pod \"dnsmasq-dns-666b6646f7-9q9hd\" (UID: \"da2af6d0-1824-4820-84eb-0c491cb777a9\") " pod="openstack/dnsmasq-dns-666b6646f7-9q9hd" Nov 29 04:53:10 crc kubenswrapper[4799]: I1129 04:53:10.607548 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da2af6d0-1824-4820-84eb-0c491cb777a9-config\") pod \"dnsmasq-dns-666b6646f7-9q9hd\" (UID: \"da2af6d0-1824-4820-84eb-0c491cb777a9\") " pod="openstack/dnsmasq-dns-666b6646f7-9q9hd" Nov 29 04:53:10 crc kubenswrapper[4799]: I1129 04:53:10.607720 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da2af6d0-1824-4820-84eb-0c491cb777a9-dns-svc\") pod \"dnsmasq-dns-666b6646f7-9q9hd\" (UID: \"da2af6d0-1824-4820-84eb-0c491cb777a9\") " pod="openstack/dnsmasq-dns-666b6646f7-9q9hd" Nov 29 04:53:10 crc kubenswrapper[4799]: I1129 04:53:10.621492 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-ccgqx"] Nov 29 04:53:10 crc kubenswrapper[4799]: I1129 04:53:10.626758 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-ccgqx" Nov 29 04:53:10 crc kubenswrapper[4799]: I1129 04:53:10.647282 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccqck\" (UniqueName: \"kubernetes.io/projected/da2af6d0-1824-4820-84eb-0c491cb777a9-kube-api-access-ccqck\") pod \"dnsmasq-dns-666b6646f7-9q9hd\" (UID: \"da2af6d0-1824-4820-84eb-0c491cb777a9\") " pod="openstack/dnsmasq-dns-666b6646f7-9q9hd" Nov 29 04:53:10 crc kubenswrapper[4799]: I1129 04:53:10.704864 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-ccgqx"] Nov 29 04:53:10 crc kubenswrapper[4799]: I1129 04:53:10.710307 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f3627f09-e94d-4a00-94a2-7dcf7dfe7800-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-ccgqx\" (UID: \"f3627f09-e94d-4a00-94a2-7dcf7dfe7800\") " pod="openstack/dnsmasq-dns-57d769cc4f-ccgqx" Nov 29 04:53:10 crc kubenswrapper[4799]: I1129 04:53:10.710665 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3627f09-e94d-4a00-94a2-7dcf7dfe7800-config\") pod \"dnsmasq-dns-57d769cc4f-ccgqx\" (UID: \"f3627f09-e94d-4a00-94a2-7dcf7dfe7800\") " pod="openstack/dnsmasq-dns-57d769cc4f-ccgqx" Nov 29 04:53:10 crc kubenswrapper[4799]: I1129 04:53:10.710758 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crcdk\" (UniqueName: \"kubernetes.io/projected/f3627f09-e94d-4a00-94a2-7dcf7dfe7800-kube-api-access-crcdk\") pod \"dnsmasq-dns-57d769cc4f-ccgqx\" (UID: \"f3627f09-e94d-4a00-94a2-7dcf7dfe7800\") " pod="openstack/dnsmasq-dns-57d769cc4f-ccgqx" Nov 29 04:53:10 crc kubenswrapper[4799]: I1129 04:53:10.814155 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f3627f09-e94d-4a00-94a2-7dcf7dfe7800-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-ccgqx\" (UID: \"f3627f09-e94d-4a00-94a2-7dcf7dfe7800\") " pod="openstack/dnsmasq-dns-57d769cc4f-ccgqx" Nov 29 04:53:10 crc kubenswrapper[4799]: I1129 04:53:10.814294 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3627f09-e94d-4a00-94a2-7dcf7dfe7800-config\") pod \"dnsmasq-dns-57d769cc4f-ccgqx\" (UID: \"f3627f09-e94d-4a00-94a2-7dcf7dfe7800\") " pod="openstack/dnsmasq-dns-57d769cc4f-ccgqx" Nov 29 04:53:10 crc kubenswrapper[4799]: I1129 04:53:10.814361 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crcdk\" (UniqueName: \"kubernetes.io/projected/f3627f09-e94d-4a00-94a2-7dcf7dfe7800-kube-api-access-crcdk\") pod \"dnsmasq-dns-57d769cc4f-ccgqx\" (UID: \"f3627f09-e94d-4a00-94a2-7dcf7dfe7800\") " pod="openstack/dnsmasq-dns-57d769cc4f-ccgqx" Nov 29 04:53:10 crc kubenswrapper[4799]: I1129 04:53:10.815712 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f3627f09-e94d-4a00-94a2-7dcf7dfe7800-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-ccgqx\" (UID: \"f3627f09-e94d-4a00-94a2-7dcf7dfe7800\") " pod="openstack/dnsmasq-dns-57d769cc4f-ccgqx" Nov 29 04:53:10 crc kubenswrapper[4799]: I1129 04:53:10.823702 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3627f09-e94d-4a00-94a2-7dcf7dfe7800-config\") pod \"dnsmasq-dns-57d769cc4f-ccgqx\" (UID: \"f3627f09-e94d-4a00-94a2-7dcf7dfe7800\") " pod="openstack/dnsmasq-dns-57d769cc4f-ccgqx" Nov 29 04:53:10 crc kubenswrapper[4799]: I1129 04:53:10.840410 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crcdk\" (UniqueName: \"kubernetes.io/projected/f3627f09-e94d-4a00-94a2-7dcf7dfe7800-kube-api-access-crcdk\") pod \"dnsmasq-dns-57d769cc4f-ccgqx\" (UID: \"f3627f09-e94d-4a00-94a2-7dcf7dfe7800\") " pod="openstack/dnsmasq-dns-57d769cc4f-ccgqx" Nov 29 04:53:10 crc kubenswrapper[4799]: I1129 04:53:10.932074 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-9q9hd" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.005257 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-ccgqx" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.425527 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.427298 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.430429 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.430436 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.430817 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-7gp88" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.430865 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.430950 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.430972 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.431052 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.454971 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.495751 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-9q9hd"] Nov 29 04:53:11 crc kubenswrapper[4799]: W1129 04:53:11.510447 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podda2af6d0_1824_4820_84eb_0c491cb777a9.slice/crio-ad8249dfae367d7e0f5593bfe8683dd96de7e32947539f1423cba28b54dbc50a WatchSource:0}: Error finding container ad8249dfae367d7e0f5593bfe8683dd96de7e32947539f1423cba28b54dbc50a: Status 404 returned error can't find the container with id ad8249dfae367d7e0f5593bfe8683dd96de7e32947539f1423cba28b54dbc50a Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.530484 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1788e4d3-8c2a-4d06-b746-c0fb12514148-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"1788e4d3-8c2a-4d06-b746-c0fb12514148\") " pod="openstack/rabbitmq-server-0" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.530623 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1788e4d3-8c2a-4d06-b746-c0fb12514148-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"1788e4d3-8c2a-4d06-b746-c0fb12514148\") " pod="openstack/rabbitmq-server-0" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.530657 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1788e4d3-8c2a-4d06-b746-c0fb12514148-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"1788e4d3-8c2a-4d06-b746-c0fb12514148\") " pod="openstack/rabbitmq-server-0" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.530714 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k94v6\" (UniqueName: \"kubernetes.io/projected/1788e4d3-8c2a-4d06-b746-c0fb12514148-kube-api-access-k94v6\") pod \"rabbitmq-server-0\" (UID: \"1788e4d3-8c2a-4d06-b746-c0fb12514148\") " pod="openstack/rabbitmq-server-0" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.530747 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1788e4d3-8c2a-4d06-b746-c0fb12514148-config-data\") pod \"rabbitmq-server-0\" (UID: \"1788e4d3-8c2a-4d06-b746-c0fb12514148\") " pod="openstack/rabbitmq-server-0" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.530898 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1788e4d3-8c2a-4d06-b746-c0fb12514148-server-conf\") pod \"rabbitmq-server-0\" (UID: \"1788e4d3-8c2a-4d06-b746-c0fb12514148\") " pod="openstack/rabbitmq-server-0" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.530936 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1788e4d3-8c2a-4d06-b746-c0fb12514148-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"1788e4d3-8c2a-4d06-b746-c0fb12514148\") " pod="openstack/rabbitmq-server-0" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.531024 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1788e4d3-8c2a-4d06-b746-c0fb12514148-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"1788e4d3-8c2a-4d06-b746-c0fb12514148\") " pod="openstack/rabbitmq-server-0" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.531084 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1788e4d3-8c2a-4d06-b746-c0fb12514148-pod-info\") pod \"rabbitmq-server-0\" (UID: \"1788e4d3-8c2a-4d06-b746-c0fb12514148\") " pod="openstack/rabbitmq-server-0" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.531150 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"1788e4d3-8c2a-4d06-b746-c0fb12514148\") " pod="openstack/rabbitmq-server-0" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.531185 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1788e4d3-8c2a-4d06-b746-c0fb12514148-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"1788e4d3-8c2a-4d06-b746-c0fb12514148\") " pod="openstack/rabbitmq-server-0" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.621132 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-ccgqx"] Nov 29 04:53:11 crc kubenswrapper[4799]: W1129 04:53:11.629009 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf3627f09_e94d_4a00_94a2_7dcf7dfe7800.slice/crio-02adf8c7c010cdf77d280123300275260a01c4c9ec438444c80d8a0c324b69a2 WatchSource:0}: Error finding container 02adf8c7c010cdf77d280123300275260a01c4c9ec438444c80d8a0c324b69a2: Status 404 returned error can't find the container with id 02adf8c7c010cdf77d280123300275260a01c4c9ec438444c80d8a0c324b69a2 Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.632252 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k94v6\" (UniqueName: \"kubernetes.io/projected/1788e4d3-8c2a-4d06-b746-c0fb12514148-kube-api-access-k94v6\") pod \"rabbitmq-server-0\" (UID: \"1788e4d3-8c2a-4d06-b746-c0fb12514148\") " pod="openstack/rabbitmq-server-0" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.632441 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1788e4d3-8c2a-4d06-b746-c0fb12514148-config-data\") pod \"rabbitmq-server-0\" (UID: \"1788e4d3-8c2a-4d06-b746-c0fb12514148\") " pod="openstack/rabbitmq-server-0" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.632516 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1788e4d3-8c2a-4d06-b746-c0fb12514148-server-conf\") pod \"rabbitmq-server-0\" (UID: \"1788e4d3-8c2a-4d06-b746-c0fb12514148\") " pod="openstack/rabbitmq-server-0" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.633665 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1788e4d3-8c2a-4d06-b746-c0fb12514148-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"1788e4d3-8c2a-4d06-b746-c0fb12514148\") " pod="openstack/rabbitmq-server-0" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.634499 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1788e4d3-8c2a-4d06-b746-c0fb12514148-config-data\") pod \"rabbitmq-server-0\" (UID: \"1788e4d3-8c2a-4d06-b746-c0fb12514148\") " pod="openstack/rabbitmq-server-0" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.634918 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1788e4d3-8c2a-4d06-b746-c0fb12514148-server-conf\") pod \"rabbitmq-server-0\" (UID: \"1788e4d3-8c2a-4d06-b746-c0fb12514148\") " pod="openstack/rabbitmq-server-0" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.634993 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1788e4d3-8c2a-4d06-b746-c0fb12514148-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"1788e4d3-8c2a-4d06-b746-c0fb12514148\") " pod="openstack/rabbitmq-server-0" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.635149 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1788e4d3-8c2a-4d06-b746-c0fb12514148-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"1788e4d3-8c2a-4d06-b746-c0fb12514148\") " pod="openstack/rabbitmq-server-0" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.635942 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1788e4d3-8c2a-4d06-b746-c0fb12514148-pod-info\") pod \"rabbitmq-server-0\" (UID: \"1788e4d3-8c2a-4d06-b746-c0fb12514148\") " pod="openstack/rabbitmq-server-0" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.636081 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"1788e4d3-8c2a-4d06-b746-c0fb12514148\") " pod="openstack/rabbitmq-server-0" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.636145 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1788e4d3-8c2a-4d06-b746-c0fb12514148-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"1788e4d3-8c2a-4d06-b746-c0fb12514148\") " pod="openstack/rabbitmq-server-0" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.636274 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1788e4d3-8c2a-4d06-b746-c0fb12514148-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"1788e4d3-8c2a-4d06-b746-c0fb12514148\") " pod="openstack/rabbitmq-server-0" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.636384 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1788e4d3-8c2a-4d06-b746-c0fb12514148-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"1788e4d3-8c2a-4d06-b746-c0fb12514148\") " pod="openstack/rabbitmq-server-0" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.636419 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1788e4d3-8c2a-4d06-b746-c0fb12514148-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"1788e4d3-8c2a-4d06-b746-c0fb12514148\") " pod="openstack/rabbitmq-server-0" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.636637 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"1788e4d3-8c2a-4d06-b746-c0fb12514148\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-server-0" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.637992 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1788e4d3-8c2a-4d06-b746-c0fb12514148-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"1788e4d3-8c2a-4d06-b746-c0fb12514148\") " pod="openstack/rabbitmq-server-0" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.639421 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1788e4d3-8c2a-4d06-b746-c0fb12514148-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"1788e4d3-8c2a-4d06-b746-c0fb12514148\") " pod="openstack/rabbitmq-server-0" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.644101 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1788e4d3-8c2a-4d06-b746-c0fb12514148-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"1788e4d3-8c2a-4d06-b746-c0fb12514148\") " pod="openstack/rabbitmq-server-0" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.644183 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1788e4d3-8c2a-4d06-b746-c0fb12514148-pod-info\") pod \"rabbitmq-server-0\" (UID: \"1788e4d3-8c2a-4d06-b746-c0fb12514148\") " pod="openstack/rabbitmq-server-0" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.645092 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1788e4d3-8c2a-4d06-b746-c0fb12514148-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"1788e4d3-8c2a-4d06-b746-c0fb12514148\") " pod="openstack/rabbitmq-server-0" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.661116 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1788e4d3-8c2a-4d06-b746-c0fb12514148-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"1788e4d3-8c2a-4d06-b746-c0fb12514148\") " pod="openstack/rabbitmq-server-0" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.664625 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k94v6\" (UniqueName: \"kubernetes.io/projected/1788e4d3-8c2a-4d06-b746-c0fb12514148-kube-api-access-k94v6\") pod \"rabbitmq-server-0\" (UID: \"1788e4d3-8c2a-4d06-b746-c0fb12514148\") " pod="openstack/rabbitmq-server-0" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.670315 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"1788e4d3-8c2a-4d06-b746-c0fb12514148\") " pod="openstack/rabbitmq-server-0" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.765630 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.776595 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.778300 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.781710 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.782197 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.782252 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.782392 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.782374 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.782697 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.782826 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-jhdhp" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.793183 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.944597 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z859g\" (UniqueName: \"kubernetes.io/projected/81726039-72e9-455a-b4fb-e16184acdc47-kube-api-access-z859g\") pod \"rabbitmq-cell1-server-0\" (UID: \"81726039-72e9-455a-b4fb-e16184acdc47\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.945049 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/81726039-72e9-455a-b4fb-e16184acdc47-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"81726039-72e9-455a-b4fb-e16184acdc47\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.945082 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/81726039-72e9-455a-b4fb-e16184acdc47-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"81726039-72e9-455a-b4fb-e16184acdc47\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.945109 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/81726039-72e9-455a-b4fb-e16184acdc47-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"81726039-72e9-455a-b4fb-e16184acdc47\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.945163 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/81726039-72e9-455a-b4fb-e16184acdc47-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"81726039-72e9-455a-b4fb-e16184acdc47\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.945273 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/81726039-72e9-455a-b4fb-e16184acdc47-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"81726039-72e9-455a-b4fb-e16184acdc47\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.945327 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"81726039-72e9-455a-b4fb-e16184acdc47\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.945357 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/81726039-72e9-455a-b4fb-e16184acdc47-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"81726039-72e9-455a-b4fb-e16184acdc47\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.945398 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/81726039-72e9-455a-b4fb-e16184acdc47-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"81726039-72e9-455a-b4fb-e16184acdc47\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.945449 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/81726039-72e9-455a-b4fb-e16184acdc47-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"81726039-72e9-455a-b4fb-e16184acdc47\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:53:11 crc kubenswrapper[4799]: I1129 04:53:11.945489 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/81726039-72e9-455a-b4fb-e16184acdc47-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"81726039-72e9-455a-b4fb-e16184acdc47\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:53:12 crc kubenswrapper[4799]: I1129 04:53:12.047215 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/81726039-72e9-455a-b4fb-e16184acdc47-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"81726039-72e9-455a-b4fb-e16184acdc47\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:53:12 crc kubenswrapper[4799]: I1129 04:53:12.047308 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/81726039-72e9-455a-b4fb-e16184acdc47-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"81726039-72e9-455a-b4fb-e16184acdc47\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:53:12 crc kubenswrapper[4799]: I1129 04:53:12.047381 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/81726039-72e9-455a-b4fb-e16184acdc47-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"81726039-72e9-455a-b4fb-e16184acdc47\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:53:12 crc kubenswrapper[4799]: I1129 04:53:12.047410 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z859g\" (UniqueName: \"kubernetes.io/projected/81726039-72e9-455a-b4fb-e16184acdc47-kube-api-access-z859g\") pod \"rabbitmq-cell1-server-0\" (UID: \"81726039-72e9-455a-b4fb-e16184acdc47\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:53:12 crc kubenswrapper[4799]: I1129 04:53:12.047461 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/81726039-72e9-455a-b4fb-e16184acdc47-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"81726039-72e9-455a-b4fb-e16184acdc47\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:53:12 crc kubenswrapper[4799]: I1129 04:53:12.047484 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/81726039-72e9-455a-b4fb-e16184acdc47-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"81726039-72e9-455a-b4fb-e16184acdc47\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:53:12 crc kubenswrapper[4799]: I1129 04:53:12.047504 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/81726039-72e9-455a-b4fb-e16184acdc47-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"81726039-72e9-455a-b4fb-e16184acdc47\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:53:12 crc kubenswrapper[4799]: I1129 04:53:12.047532 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/81726039-72e9-455a-b4fb-e16184acdc47-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"81726039-72e9-455a-b4fb-e16184acdc47\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:53:12 crc kubenswrapper[4799]: I1129 04:53:12.047572 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/81726039-72e9-455a-b4fb-e16184acdc47-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"81726039-72e9-455a-b4fb-e16184acdc47\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:53:12 crc kubenswrapper[4799]: I1129 04:53:12.047597 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"81726039-72e9-455a-b4fb-e16184acdc47\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:53:12 crc kubenswrapper[4799]: I1129 04:53:12.047617 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/81726039-72e9-455a-b4fb-e16184acdc47-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"81726039-72e9-455a-b4fb-e16184acdc47\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:53:12 crc kubenswrapper[4799]: I1129 04:53:12.048488 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/81726039-72e9-455a-b4fb-e16184acdc47-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"81726039-72e9-455a-b4fb-e16184acdc47\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:53:12 crc kubenswrapper[4799]: I1129 04:53:12.048865 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"81726039-72e9-455a-b4fb-e16184acdc47\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:53:12 crc kubenswrapper[4799]: I1129 04:53:12.049435 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/81726039-72e9-455a-b4fb-e16184acdc47-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"81726039-72e9-455a-b4fb-e16184acdc47\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:53:12 crc kubenswrapper[4799]: I1129 04:53:12.049668 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/81726039-72e9-455a-b4fb-e16184acdc47-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"81726039-72e9-455a-b4fb-e16184acdc47\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:53:12 crc kubenswrapper[4799]: I1129 04:53:12.050256 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/81726039-72e9-455a-b4fb-e16184acdc47-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"81726039-72e9-455a-b4fb-e16184acdc47\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:53:12 crc kubenswrapper[4799]: I1129 04:53:12.052702 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/81726039-72e9-455a-b4fb-e16184acdc47-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"81726039-72e9-455a-b4fb-e16184acdc47\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:53:12 crc kubenswrapper[4799]: I1129 04:53:12.053692 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/81726039-72e9-455a-b4fb-e16184acdc47-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"81726039-72e9-455a-b4fb-e16184acdc47\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:53:12 crc kubenswrapper[4799]: I1129 04:53:12.058040 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/81726039-72e9-455a-b4fb-e16184acdc47-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"81726039-72e9-455a-b4fb-e16184acdc47\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:53:12 crc kubenswrapper[4799]: I1129 04:53:12.059326 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/81726039-72e9-455a-b4fb-e16184acdc47-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"81726039-72e9-455a-b4fb-e16184acdc47\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:53:12 crc kubenswrapper[4799]: I1129 04:53:12.068417 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z859g\" (UniqueName: \"kubernetes.io/projected/81726039-72e9-455a-b4fb-e16184acdc47-kube-api-access-z859g\") pod \"rabbitmq-cell1-server-0\" (UID: \"81726039-72e9-455a-b4fb-e16184acdc47\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:53:12 crc kubenswrapper[4799]: I1129 04:53:12.073048 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/81726039-72e9-455a-b4fb-e16184acdc47-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"81726039-72e9-455a-b4fb-e16184acdc47\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:53:12 crc kubenswrapper[4799]: I1129 04:53:12.080300 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"81726039-72e9-455a-b4fb-e16184acdc47\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:53:12 crc kubenswrapper[4799]: I1129 04:53:12.107436 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:53:12 crc kubenswrapper[4799]: I1129 04:53:12.230720 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-9q9hd" event={"ID":"da2af6d0-1824-4820-84eb-0c491cb777a9","Type":"ContainerStarted","Data":"ad8249dfae367d7e0f5593bfe8683dd96de7e32947539f1423cba28b54dbc50a"} Nov 29 04:53:12 crc kubenswrapper[4799]: I1129 04:53:12.233141 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-ccgqx" event={"ID":"f3627f09-e94d-4a00-94a2-7dcf7dfe7800","Type":"ContainerStarted","Data":"02adf8c7c010cdf77d280123300275260a01c4c9ec438444c80d8a0c324b69a2"} Nov 29 04:53:12 crc kubenswrapper[4799]: I1129 04:53:12.372811 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 29 04:53:12 crc kubenswrapper[4799]: W1129 04:53:12.377866 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1788e4d3_8c2a_4d06_b746_c0fb12514148.slice/crio-25b331b3efc2702cb80297c00783deea910cc25e6c427e635005bcaff6f39e61 WatchSource:0}: Error finding container 25b331b3efc2702cb80297c00783deea910cc25e6c427e635005bcaff6f39e61: Status 404 returned error can't find the container with id 25b331b3efc2702cb80297c00783deea910cc25e6c427e635005bcaff6f39e61 Nov 29 04:53:12 crc kubenswrapper[4799]: I1129 04:53:12.688369 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 29 04:53:12 crc kubenswrapper[4799]: W1129 04:53:12.693886 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod81726039_72e9_455a_b4fb_e16184acdc47.slice/crio-3bfd2d2b4232ba9982c50d11901772e1aed4dcbdae556a40673d8ea38d646aa1 WatchSource:0}: Error finding container 3bfd2d2b4232ba9982c50d11901772e1aed4dcbdae556a40673d8ea38d646aa1: Status 404 returned error can't find the container with id 3bfd2d2b4232ba9982c50d11901772e1aed4dcbdae556a40673d8ea38d646aa1 Nov 29 04:53:13 crc kubenswrapper[4799]: I1129 04:53:13.245905 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"81726039-72e9-455a-b4fb-e16184acdc47","Type":"ContainerStarted","Data":"3bfd2d2b4232ba9982c50d11901772e1aed4dcbdae556a40673d8ea38d646aa1"} Nov 29 04:53:13 crc kubenswrapper[4799]: I1129 04:53:13.248956 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1788e4d3-8c2a-4d06-b746-c0fb12514148","Type":"ContainerStarted","Data":"25b331b3efc2702cb80297c00783deea910cc25e6c427e635005bcaff6f39e61"} Nov 29 04:53:13 crc kubenswrapper[4799]: I1129 04:53:13.481835 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Nov 29 04:53:13 crc kubenswrapper[4799]: I1129 04:53:13.485061 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 29 04:53:13 crc kubenswrapper[4799]: I1129 04:53:13.488923 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-xzswh" Nov 29 04:53:13 crc kubenswrapper[4799]: I1129 04:53:13.489775 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Nov 29 04:53:13 crc kubenswrapper[4799]: I1129 04:53:13.489808 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Nov 29 04:53:13 crc kubenswrapper[4799]: I1129 04:53:13.490206 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Nov 29 04:53:13 crc kubenswrapper[4799]: I1129 04:53:13.493802 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 29 04:53:13 crc kubenswrapper[4799]: I1129 04:53:13.496436 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Nov 29 04:53:13 crc kubenswrapper[4799]: I1129 04:53:13.581610 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbc1f3a5-585d-4e6d-a87e-cf77a7b21147-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"bbc1f3a5-585d-4e6d-a87e-cf77a7b21147\") " pod="openstack/openstack-galera-0" Nov 29 04:53:13 crc kubenswrapper[4799]: I1129 04:53:13.582394 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bbc1f3a5-585d-4e6d-a87e-cf77a7b21147-operator-scripts\") pod \"openstack-galera-0\" (UID: \"bbc1f3a5-585d-4e6d-a87e-cf77a7b21147\") " pod="openstack/openstack-galera-0" Nov 29 04:53:13 crc kubenswrapper[4799]: I1129 04:53:13.582472 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4v9bj\" (UniqueName: \"kubernetes.io/projected/bbc1f3a5-585d-4e6d-a87e-cf77a7b21147-kube-api-access-4v9bj\") pod \"openstack-galera-0\" (UID: \"bbc1f3a5-585d-4e6d-a87e-cf77a7b21147\") " pod="openstack/openstack-galera-0" Nov 29 04:53:13 crc kubenswrapper[4799]: I1129 04:53:13.582522 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/bbc1f3a5-585d-4e6d-a87e-cf77a7b21147-kolla-config\") pod \"openstack-galera-0\" (UID: \"bbc1f3a5-585d-4e6d-a87e-cf77a7b21147\") " pod="openstack/openstack-galera-0" Nov 29 04:53:13 crc kubenswrapper[4799]: I1129 04:53:13.582562 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/bbc1f3a5-585d-4e6d-a87e-cf77a7b21147-config-data-default\") pod \"openstack-galera-0\" (UID: \"bbc1f3a5-585d-4e6d-a87e-cf77a7b21147\") " pod="openstack/openstack-galera-0" Nov 29 04:53:13 crc kubenswrapper[4799]: I1129 04:53:13.582639 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/bbc1f3a5-585d-4e6d-a87e-cf77a7b21147-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"bbc1f3a5-585d-4e6d-a87e-cf77a7b21147\") " pod="openstack/openstack-galera-0" Nov 29 04:53:13 crc kubenswrapper[4799]: I1129 04:53:13.582679 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/bbc1f3a5-585d-4e6d-a87e-cf77a7b21147-config-data-generated\") pod \"openstack-galera-0\" (UID: \"bbc1f3a5-585d-4e6d-a87e-cf77a7b21147\") " pod="openstack/openstack-galera-0" Nov 29 04:53:13 crc kubenswrapper[4799]: I1129 04:53:13.582745 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"bbc1f3a5-585d-4e6d-a87e-cf77a7b21147\") " pod="openstack/openstack-galera-0" Nov 29 04:53:13 crc kubenswrapper[4799]: I1129 04:53:13.687930 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbc1f3a5-585d-4e6d-a87e-cf77a7b21147-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"bbc1f3a5-585d-4e6d-a87e-cf77a7b21147\") " pod="openstack/openstack-galera-0" Nov 29 04:53:13 crc kubenswrapper[4799]: I1129 04:53:13.708951 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bbc1f3a5-585d-4e6d-a87e-cf77a7b21147-operator-scripts\") pod \"openstack-galera-0\" (UID: \"bbc1f3a5-585d-4e6d-a87e-cf77a7b21147\") " pod="openstack/openstack-galera-0" Nov 29 04:53:13 crc kubenswrapper[4799]: I1129 04:53:13.709130 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4v9bj\" (UniqueName: \"kubernetes.io/projected/bbc1f3a5-585d-4e6d-a87e-cf77a7b21147-kube-api-access-4v9bj\") pod \"openstack-galera-0\" (UID: \"bbc1f3a5-585d-4e6d-a87e-cf77a7b21147\") " pod="openstack/openstack-galera-0" Nov 29 04:53:13 crc kubenswrapper[4799]: I1129 04:53:13.713193 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/bbc1f3a5-585d-4e6d-a87e-cf77a7b21147-kolla-config\") pod \"openstack-galera-0\" (UID: \"bbc1f3a5-585d-4e6d-a87e-cf77a7b21147\") " pod="openstack/openstack-galera-0" Nov 29 04:53:13 crc kubenswrapper[4799]: I1129 04:53:13.713942 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/bbc1f3a5-585d-4e6d-a87e-cf77a7b21147-kolla-config\") pod \"openstack-galera-0\" (UID: \"bbc1f3a5-585d-4e6d-a87e-cf77a7b21147\") " pod="openstack/openstack-galera-0" Nov 29 04:53:13 crc kubenswrapper[4799]: I1129 04:53:13.714211 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/bbc1f3a5-585d-4e6d-a87e-cf77a7b21147-config-data-default\") pod \"openstack-galera-0\" (UID: \"bbc1f3a5-585d-4e6d-a87e-cf77a7b21147\") " pod="openstack/openstack-galera-0" Nov 29 04:53:13 crc kubenswrapper[4799]: I1129 04:53:13.714269 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/bbc1f3a5-585d-4e6d-a87e-cf77a7b21147-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"bbc1f3a5-585d-4e6d-a87e-cf77a7b21147\") " pod="openstack/openstack-galera-0" Nov 29 04:53:13 crc kubenswrapper[4799]: I1129 04:53:13.714291 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/bbc1f3a5-585d-4e6d-a87e-cf77a7b21147-config-data-generated\") pod \"openstack-galera-0\" (UID: \"bbc1f3a5-585d-4e6d-a87e-cf77a7b21147\") " pod="openstack/openstack-galera-0" Nov 29 04:53:13 crc kubenswrapper[4799]: I1129 04:53:13.715108 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/bbc1f3a5-585d-4e6d-a87e-cf77a7b21147-config-data-generated\") pod \"openstack-galera-0\" (UID: \"bbc1f3a5-585d-4e6d-a87e-cf77a7b21147\") " pod="openstack/openstack-galera-0" Nov 29 04:53:13 crc kubenswrapper[4799]: I1129 04:53:13.715129 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bbc1f3a5-585d-4e6d-a87e-cf77a7b21147-operator-scripts\") pod \"openstack-galera-0\" (UID: \"bbc1f3a5-585d-4e6d-a87e-cf77a7b21147\") " pod="openstack/openstack-galera-0" Nov 29 04:53:13 crc kubenswrapper[4799]: I1129 04:53:13.715511 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"bbc1f3a5-585d-4e6d-a87e-cf77a7b21147\") " pod="openstack/openstack-galera-0" Nov 29 04:53:13 crc kubenswrapper[4799]: I1129 04:53:13.717700 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/bbc1f3a5-585d-4e6d-a87e-cf77a7b21147-config-data-default\") pod \"openstack-galera-0\" (UID: \"bbc1f3a5-585d-4e6d-a87e-cf77a7b21147\") " pod="openstack/openstack-galera-0" Nov 29 04:53:13 crc kubenswrapper[4799]: I1129 04:53:13.720951 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"bbc1f3a5-585d-4e6d-a87e-cf77a7b21147\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/openstack-galera-0" Nov 29 04:53:13 crc kubenswrapper[4799]: I1129 04:53:13.726589 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbc1f3a5-585d-4e6d-a87e-cf77a7b21147-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"bbc1f3a5-585d-4e6d-a87e-cf77a7b21147\") " pod="openstack/openstack-galera-0" Nov 29 04:53:13 crc kubenswrapper[4799]: I1129 04:53:13.734196 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/bbc1f3a5-585d-4e6d-a87e-cf77a7b21147-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"bbc1f3a5-585d-4e6d-a87e-cf77a7b21147\") " pod="openstack/openstack-galera-0" Nov 29 04:53:13 crc kubenswrapper[4799]: I1129 04:53:13.739876 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4v9bj\" (UniqueName: \"kubernetes.io/projected/bbc1f3a5-585d-4e6d-a87e-cf77a7b21147-kube-api-access-4v9bj\") pod \"openstack-galera-0\" (UID: \"bbc1f3a5-585d-4e6d-a87e-cf77a7b21147\") " pod="openstack/openstack-galera-0" Nov 29 04:53:13 crc kubenswrapper[4799]: I1129 04:53:13.770892 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"bbc1f3a5-585d-4e6d-a87e-cf77a7b21147\") " pod="openstack/openstack-galera-0" Nov 29 04:53:13 crc kubenswrapper[4799]: I1129 04:53:13.813223 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 29 04:53:14 crc kubenswrapper[4799]: I1129 04:53:14.528184 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 29 04:53:14 crc kubenswrapper[4799]: I1129 04:53:14.912836 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 29 04:53:14 crc kubenswrapper[4799]: I1129 04:53:14.915457 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 29 04:53:14 crc kubenswrapper[4799]: I1129 04:53:14.919320 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-spw8x" Nov 29 04:53:14 crc kubenswrapper[4799]: I1129 04:53:14.919421 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Nov 29 04:53:14 crc kubenswrapper[4799]: I1129 04:53:14.919627 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Nov 29 04:53:14 crc kubenswrapper[4799]: I1129 04:53:14.921578 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Nov 29 04:53:14 crc kubenswrapper[4799]: I1129 04:53:14.924209 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 29 04:53:15 crc kubenswrapper[4799]: I1129 04:53:15.050237 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/735902f6-652e-4c85-818d-10f7d9529788-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"735902f6-652e-4c85-818d-10f7d9529788\") " pod="openstack/openstack-cell1-galera-0" Nov 29 04:53:15 crc kubenswrapper[4799]: I1129 04:53:15.050323 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/735902f6-652e-4c85-818d-10f7d9529788-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"735902f6-652e-4c85-818d-10f7d9529788\") " pod="openstack/openstack-cell1-galera-0" Nov 29 04:53:15 crc kubenswrapper[4799]: I1129 04:53:15.050464 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/735902f6-652e-4c85-818d-10f7d9529788-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"735902f6-652e-4c85-818d-10f7d9529788\") " pod="openstack/openstack-cell1-galera-0" Nov 29 04:53:15 crc kubenswrapper[4799]: I1129 04:53:15.050497 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5jkr\" (UniqueName: \"kubernetes.io/projected/735902f6-652e-4c85-818d-10f7d9529788-kube-api-access-j5jkr\") pod \"openstack-cell1-galera-0\" (UID: \"735902f6-652e-4c85-818d-10f7d9529788\") " pod="openstack/openstack-cell1-galera-0" Nov 29 04:53:15 crc kubenswrapper[4799]: I1129 04:53:15.050533 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/735902f6-652e-4c85-818d-10f7d9529788-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"735902f6-652e-4c85-818d-10f7d9529788\") " pod="openstack/openstack-cell1-galera-0" Nov 29 04:53:15 crc kubenswrapper[4799]: I1129 04:53:15.050570 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/735902f6-652e-4c85-818d-10f7d9529788-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"735902f6-652e-4c85-818d-10f7d9529788\") " pod="openstack/openstack-cell1-galera-0" Nov 29 04:53:15 crc kubenswrapper[4799]: I1129 04:53:15.050603 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/735902f6-652e-4c85-818d-10f7d9529788-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"735902f6-652e-4c85-818d-10f7d9529788\") " pod="openstack/openstack-cell1-galera-0" Nov 29 04:53:15 crc kubenswrapper[4799]: I1129 04:53:15.050627 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"735902f6-652e-4c85-818d-10f7d9529788\") " pod="openstack/openstack-cell1-galera-0" Nov 29 04:53:15 crc kubenswrapper[4799]: I1129 04:53:15.057741 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Nov 29 04:53:15 crc kubenswrapper[4799]: I1129 04:53:15.059017 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 29 04:53:15 crc kubenswrapper[4799]: I1129 04:53:15.064981 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Nov 29 04:53:15 crc kubenswrapper[4799]: I1129 04:53:15.066092 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-dk69z" Nov 29 04:53:15 crc kubenswrapper[4799]: I1129 04:53:15.067121 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Nov 29 04:53:15 crc kubenswrapper[4799]: I1129 04:53:15.092544 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 29 04:53:15 crc kubenswrapper[4799]: I1129 04:53:15.152512 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6f5ef6d-bec8-4b82-8ebc-53952954134c-combined-ca-bundle\") pod \"memcached-0\" (UID: \"b6f5ef6d-bec8-4b82-8ebc-53952954134c\") " pod="openstack/memcached-0" Nov 29 04:53:15 crc kubenswrapper[4799]: I1129 04:53:15.152874 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/735902f6-652e-4c85-818d-10f7d9529788-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"735902f6-652e-4c85-818d-10f7d9529788\") " pod="openstack/openstack-cell1-galera-0" Nov 29 04:53:15 crc kubenswrapper[4799]: I1129 04:53:15.152908 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5jkr\" (UniqueName: \"kubernetes.io/projected/735902f6-652e-4c85-818d-10f7d9529788-kube-api-access-j5jkr\") pod \"openstack-cell1-galera-0\" (UID: \"735902f6-652e-4c85-818d-10f7d9529788\") " pod="openstack/openstack-cell1-galera-0" Nov 29 04:53:15 crc kubenswrapper[4799]: I1129 04:53:15.152939 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwzxj\" (UniqueName: \"kubernetes.io/projected/b6f5ef6d-bec8-4b82-8ebc-53952954134c-kube-api-access-kwzxj\") pod \"memcached-0\" (UID: \"b6f5ef6d-bec8-4b82-8ebc-53952954134c\") " pod="openstack/memcached-0" Nov 29 04:53:15 crc kubenswrapper[4799]: I1129 04:53:15.152960 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/735902f6-652e-4c85-818d-10f7d9529788-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"735902f6-652e-4c85-818d-10f7d9529788\") " pod="openstack/openstack-cell1-galera-0" Nov 29 04:53:15 crc kubenswrapper[4799]: I1129 04:53:15.152991 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/735902f6-652e-4c85-818d-10f7d9529788-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"735902f6-652e-4c85-818d-10f7d9529788\") " pod="openstack/openstack-cell1-galera-0" Nov 29 04:53:15 crc kubenswrapper[4799]: I1129 04:53:15.153020 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/735902f6-652e-4c85-818d-10f7d9529788-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"735902f6-652e-4c85-818d-10f7d9529788\") " pod="openstack/openstack-cell1-galera-0" Nov 29 04:53:15 crc kubenswrapper[4799]: I1129 04:53:15.153040 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"735902f6-652e-4c85-818d-10f7d9529788\") " pod="openstack/openstack-cell1-galera-0" Nov 29 04:53:15 crc kubenswrapper[4799]: I1129 04:53:15.153077 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/735902f6-652e-4c85-818d-10f7d9529788-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"735902f6-652e-4c85-818d-10f7d9529788\") " pod="openstack/openstack-cell1-galera-0" Nov 29 04:53:15 crc kubenswrapper[4799]: I1129 04:53:15.153103 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b6f5ef6d-bec8-4b82-8ebc-53952954134c-kolla-config\") pod \"memcached-0\" (UID: \"b6f5ef6d-bec8-4b82-8ebc-53952954134c\") " pod="openstack/memcached-0" Nov 29 04:53:15 crc kubenswrapper[4799]: I1129 04:53:15.153137 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/735902f6-652e-4c85-818d-10f7d9529788-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"735902f6-652e-4c85-818d-10f7d9529788\") " pod="openstack/openstack-cell1-galera-0" Nov 29 04:53:15 crc kubenswrapper[4799]: I1129 04:53:15.153184 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b6f5ef6d-bec8-4b82-8ebc-53952954134c-config-data\") pod \"memcached-0\" (UID: \"b6f5ef6d-bec8-4b82-8ebc-53952954134c\") " pod="openstack/memcached-0" Nov 29 04:53:15 crc kubenswrapper[4799]: I1129 04:53:15.153217 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6f5ef6d-bec8-4b82-8ebc-53952954134c-memcached-tls-certs\") pod \"memcached-0\" (UID: \"b6f5ef6d-bec8-4b82-8ebc-53952954134c\") " pod="openstack/memcached-0" Nov 29 04:53:15 crc kubenswrapper[4799]: I1129 04:53:15.153334 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/735902f6-652e-4c85-818d-10f7d9529788-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"735902f6-652e-4c85-818d-10f7d9529788\") " pod="openstack/openstack-cell1-galera-0" Nov 29 04:53:15 crc kubenswrapper[4799]: I1129 04:53:15.153705 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"735902f6-652e-4c85-818d-10f7d9529788\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/openstack-cell1-galera-0" Nov 29 04:53:15 crc kubenswrapper[4799]: I1129 04:53:15.154956 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/735902f6-652e-4c85-818d-10f7d9529788-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"735902f6-652e-4c85-818d-10f7d9529788\") " pod="openstack/openstack-cell1-galera-0" Nov 29 04:53:15 crc kubenswrapper[4799]: I1129 04:53:15.156621 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/735902f6-652e-4c85-818d-10f7d9529788-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"735902f6-652e-4c85-818d-10f7d9529788\") " pod="openstack/openstack-cell1-galera-0" Nov 29 04:53:15 crc kubenswrapper[4799]: I1129 04:53:15.157800 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/735902f6-652e-4c85-818d-10f7d9529788-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"735902f6-652e-4c85-818d-10f7d9529788\") " pod="openstack/openstack-cell1-galera-0" Nov 29 04:53:15 crc kubenswrapper[4799]: I1129 04:53:15.162051 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/735902f6-652e-4c85-818d-10f7d9529788-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"735902f6-652e-4c85-818d-10f7d9529788\") " pod="openstack/openstack-cell1-galera-0" Nov 29 04:53:15 crc kubenswrapper[4799]: I1129 04:53:15.169295 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/735902f6-652e-4c85-818d-10f7d9529788-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"735902f6-652e-4c85-818d-10f7d9529788\") " pod="openstack/openstack-cell1-galera-0" Nov 29 04:53:15 crc kubenswrapper[4799]: I1129 04:53:15.176115 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5jkr\" (UniqueName: \"kubernetes.io/projected/735902f6-652e-4c85-818d-10f7d9529788-kube-api-access-j5jkr\") pod \"openstack-cell1-galera-0\" (UID: \"735902f6-652e-4c85-818d-10f7d9529788\") " pod="openstack/openstack-cell1-galera-0" Nov 29 04:53:15 crc kubenswrapper[4799]: I1129 04:53:15.189722 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"735902f6-652e-4c85-818d-10f7d9529788\") " pod="openstack/openstack-cell1-galera-0" Nov 29 04:53:15 crc kubenswrapper[4799]: I1129 04:53:15.251575 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 29 04:53:15 crc kubenswrapper[4799]: I1129 04:53:15.254409 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b6f5ef6d-bec8-4b82-8ebc-53952954134c-kolla-config\") pod \"memcached-0\" (UID: \"b6f5ef6d-bec8-4b82-8ebc-53952954134c\") " pod="openstack/memcached-0" Nov 29 04:53:15 crc kubenswrapper[4799]: I1129 04:53:15.254492 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b6f5ef6d-bec8-4b82-8ebc-53952954134c-config-data\") pod \"memcached-0\" (UID: \"b6f5ef6d-bec8-4b82-8ebc-53952954134c\") " pod="openstack/memcached-0" Nov 29 04:53:15 crc kubenswrapper[4799]: I1129 04:53:15.254520 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6f5ef6d-bec8-4b82-8ebc-53952954134c-memcached-tls-certs\") pod \"memcached-0\" (UID: \"b6f5ef6d-bec8-4b82-8ebc-53952954134c\") " pod="openstack/memcached-0" Nov 29 04:53:15 crc kubenswrapper[4799]: I1129 04:53:15.254603 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6f5ef6d-bec8-4b82-8ebc-53952954134c-combined-ca-bundle\") pod \"memcached-0\" (UID: \"b6f5ef6d-bec8-4b82-8ebc-53952954134c\") " pod="openstack/memcached-0" Nov 29 04:53:15 crc kubenswrapper[4799]: I1129 04:53:15.254653 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwzxj\" (UniqueName: \"kubernetes.io/projected/b6f5ef6d-bec8-4b82-8ebc-53952954134c-kube-api-access-kwzxj\") pod \"memcached-0\" (UID: \"b6f5ef6d-bec8-4b82-8ebc-53952954134c\") " pod="openstack/memcached-0" Nov 29 04:53:15 crc kubenswrapper[4799]: I1129 04:53:15.258813 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b6f5ef6d-bec8-4b82-8ebc-53952954134c-kolla-config\") pod \"memcached-0\" (UID: \"b6f5ef6d-bec8-4b82-8ebc-53952954134c\") " pod="openstack/memcached-0" Nov 29 04:53:15 crc kubenswrapper[4799]: I1129 04:53:15.261682 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b6f5ef6d-bec8-4b82-8ebc-53952954134c-config-data\") pod \"memcached-0\" (UID: \"b6f5ef6d-bec8-4b82-8ebc-53952954134c\") " pod="openstack/memcached-0" Nov 29 04:53:15 crc kubenswrapper[4799]: I1129 04:53:15.262704 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6f5ef6d-bec8-4b82-8ebc-53952954134c-memcached-tls-certs\") pod \"memcached-0\" (UID: \"b6f5ef6d-bec8-4b82-8ebc-53952954134c\") " pod="openstack/memcached-0" Nov 29 04:53:15 crc kubenswrapper[4799]: I1129 04:53:15.269295 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6f5ef6d-bec8-4b82-8ebc-53952954134c-combined-ca-bundle\") pod \"memcached-0\" (UID: \"b6f5ef6d-bec8-4b82-8ebc-53952954134c\") " pod="openstack/memcached-0" Nov 29 04:53:15 crc kubenswrapper[4799]: I1129 04:53:15.292294 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwzxj\" (UniqueName: \"kubernetes.io/projected/b6f5ef6d-bec8-4b82-8ebc-53952954134c-kube-api-access-kwzxj\") pod \"memcached-0\" (UID: \"b6f5ef6d-bec8-4b82-8ebc-53952954134c\") " pod="openstack/memcached-0" Nov 29 04:53:15 crc kubenswrapper[4799]: I1129 04:53:15.380840 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 29 04:53:16 crc kubenswrapper[4799]: I1129 04:53:16.869619 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 29 04:53:16 crc kubenswrapper[4799]: I1129 04:53:16.871114 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 29 04:53:16 crc kubenswrapper[4799]: I1129 04:53:16.874711 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-4kr47" Nov 29 04:53:16 crc kubenswrapper[4799]: I1129 04:53:16.896027 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77k8h\" (UniqueName: \"kubernetes.io/projected/9c406f33-9e26-4226-b6dd-c7449fa96b9f-kube-api-access-77k8h\") pod \"kube-state-metrics-0\" (UID: \"9c406f33-9e26-4226-b6dd-c7449fa96b9f\") " pod="openstack/kube-state-metrics-0" Nov 29 04:53:16 crc kubenswrapper[4799]: I1129 04:53:16.981441 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 29 04:53:16 crc kubenswrapper[4799]: I1129 04:53:16.997995 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77k8h\" (UniqueName: \"kubernetes.io/projected/9c406f33-9e26-4226-b6dd-c7449fa96b9f-kube-api-access-77k8h\") pod \"kube-state-metrics-0\" (UID: \"9c406f33-9e26-4226-b6dd-c7449fa96b9f\") " pod="openstack/kube-state-metrics-0" Nov 29 04:53:17 crc kubenswrapper[4799]: I1129 04:53:17.038756 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77k8h\" (UniqueName: \"kubernetes.io/projected/9c406f33-9e26-4226-b6dd-c7449fa96b9f-kube-api-access-77k8h\") pod \"kube-state-metrics-0\" (UID: \"9c406f33-9e26-4226-b6dd-c7449fa96b9f\") " pod="openstack/kube-state-metrics-0" Nov 29 04:53:17 crc kubenswrapper[4799]: I1129 04:53:17.192203 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 29 04:53:18 crc kubenswrapper[4799]: I1129 04:53:18.985434 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7gtsn"] Nov 29 04:53:18 crc kubenswrapper[4799]: I1129 04:53:18.988337 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7gtsn" Nov 29 04:53:19 crc kubenswrapper[4799]: I1129 04:53:19.005251 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7gtsn"] Nov 29 04:53:19 crc kubenswrapper[4799]: I1129 04:53:19.040438 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80a2712a-3766-49b1-bab4-a806bb44e21f-utilities\") pod \"certified-operators-7gtsn\" (UID: \"80a2712a-3766-49b1-bab4-a806bb44e21f\") " pod="openshift-marketplace/certified-operators-7gtsn" Nov 29 04:53:19 crc kubenswrapper[4799]: I1129 04:53:19.040500 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80a2712a-3766-49b1-bab4-a806bb44e21f-catalog-content\") pod \"certified-operators-7gtsn\" (UID: \"80a2712a-3766-49b1-bab4-a806bb44e21f\") " pod="openshift-marketplace/certified-operators-7gtsn" Nov 29 04:53:19 crc kubenswrapper[4799]: I1129 04:53:19.040600 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5gfl\" (UniqueName: \"kubernetes.io/projected/80a2712a-3766-49b1-bab4-a806bb44e21f-kube-api-access-n5gfl\") pod \"certified-operators-7gtsn\" (UID: \"80a2712a-3766-49b1-bab4-a806bb44e21f\") " pod="openshift-marketplace/certified-operators-7gtsn" Nov 29 04:53:19 crc kubenswrapper[4799]: I1129 04:53:19.144817 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80a2712a-3766-49b1-bab4-a806bb44e21f-utilities\") pod \"certified-operators-7gtsn\" (UID: \"80a2712a-3766-49b1-bab4-a806bb44e21f\") " pod="openshift-marketplace/certified-operators-7gtsn" Nov 29 04:53:19 crc kubenswrapper[4799]: I1129 04:53:19.144876 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80a2712a-3766-49b1-bab4-a806bb44e21f-catalog-content\") pod \"certified-operators-7gtsn\" (UID: \"80a2712a-3766-49b1-bab4-a806bb44e21f\") " pod="openshift-marketplace/certified-operators-7gtsn" Nov 29 04:53:19 crc kubenswrapper[4799]: I1129 04:53:19.144940 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5gfl\" (UniqueName: \"kubernetes.io/projected/80a2712a-3766-49b1-bab4-a806bb44e21f-kube-api-access-n5gfl\") pod \"certified-operators-7gtsn\" (UID: \"80a2712a-3766-49b1-bab4-a806bb44e21f\") " pod="openshift-marketplace/certified-operators-7gtsn" Nov 29 04:53:19 crc kubenswrapper[4799]: I1129 04:53:19.145679 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80a2712a-3766-49b1-bab4-a806bb44e21f-utilities\") pod \"certified-operators-7gtsn\" (UID: \"80a2712a-3766-49b1-bab4-a806bb44e21f\") " pod="openshift-marketplace/certified-operators-7gtsn" Nov 29 04:53:19 crc kubenswrapper[4799]: I1129 04:53:19.145727 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80a2712a-3766-49b1-bab4-a806bb44e21f-catalog-content\") pod \"certified-operators-7gtsn\" (UID: \"80a2712a-3766-49b1-bab4-a806bb44e21f\") " pod="openshift-marketplace/certified-operators-7gtsn" Nov 29 04:53:19 crc kubenswrapper[4799]: I1129 04:53:19.177759 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5gfl\" (UniqueName: \"kubernetes.io/projected/80a2712a-3766-49b1-bab4-a806bb44e21f-kube-api-access-n5gfl\") pod \"certified-operators-7gtsn\" (UID: \"80a2712a-3766-49b1-bab4-a806bb44e21f\") " pod="openshift-marketplace/certified-operators-7gtsn" Nov 29 04:53:19 crc kubenswrapper[4799]: I1129 04:53:19.322304 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7gtsn" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.152989 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-kkdw8"] Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.154962 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-kkdw8" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.158165 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.158405 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.160406 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-kz9np" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.178882 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-kkdw8"] Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.181639 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1d3d81e5-591a-469d-9851-86271162d455-var-log-ovn\") pod \"ovn-controller-kkdw8\" (UID: \"1d3d81e5-591a-469d-9851-86271162d455\") " pod="openstack/ovn-controller-kkdw8" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.181704 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d3d81e5-591a-469d-9851-86271162d455-combined-ca-bundle\") pod \"ovn-controller-kkdw8\" (UID: \"1d3d81e5-591a-469d-9851-86271162d455\") " pod="openstack/ovn-controller-kkdw8" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.181739 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1d3d81e5-591a-469d-9851-86271162d455-var-run\") pod \"ovn-controller-kkdw8\" (UID: \"1d3d81e5-591a-469d-9851-86271162d455\") " pod="openstack/ovn-controller-kkdw8" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.181779 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d3d81e5-591a-469d-9851-86271162d455-ovn-controller-tls-certs\") pod \"ovn-controller-kkdw8\" (UID: \"1d3d81e5-591a-469d-9851-86271162d455\") " pod="openstack/ovn-controller-kkdw8" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.181857 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x245m\" (UniqueName: \"kubernetes.io/projected/1d3d81e5-591a-469d-9851-86271162d455-kube-api-access-x245m\") pod \"ovn-controller-kkdw8\" (UID: \"1d3d81e5-591a-469d-9851-86271162d455\") " pod="openstack/ovn-controller-kkdw8" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.181884 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1d3d81e5-591a-469d-9851-86271162d455-var-run-ovn\") pod \"ovn-controller-kkdw8\" (UID: \"1d3d81e5-591a-469d-9851-86271162d455\") " pod="openstack/ovn-controller-kkdw8" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.181923 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1d3d81e5-591a-469d-9851-86271162d455-scripts\") pod \"ovn-controller-kkdw8\" (UID: \"1d3d81e5-591a-469d-9851-86271162d455\") " pod="openstack/ovn-controller-kkdw8" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.250034 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-bkkst"] Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.251804 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-bkkst" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.272519 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-bkkst"] Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.283672 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/e358b742-108b-4f60-989f-2e8eddce487e-etc-ovs\") pod \"ovn-controller-ovs-bkkst\" (UID: \"e358b742-108b-4f60-989f-2e8eddce487e\") " pod="openstack/ovn-controller-ovs-bkkst" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.283759 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e358b742-108b-4f60-989f-2e8eddce487e-var-run\") pod \"ovn-controller-ovs-bkkst\" (UID: \"e358b742-108b-4f60-989f-2e8eddce487e\") " pod="openstack/ovn-controller-ovs-bkkst" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.283811 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/e358b742-108b-4f60-989f-2e8eddce487e-var-log\") pod \"ovn-controller-ovs-bkkst\" (UID: \"e358b742-108b-4f60-989f-2e8eddce487e\") " pod="openstack/ovn-controller-ovs-bkkst" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.283840 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1d3d81e5-591a-469d-9851-86271162d455-var-log-ovn\") pod \"ovn-controller-kkdw8\" (UID: \"1d3d81e5-591a-469d-9851-86271162d455\") " pod="openstack/ovn-controller-kkdw8" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.283884 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d3d81e5-591a-469d-9851-86271162d455-combined-ca-bundle\") pod \"ovn-controller-kkdw8\" (UID: \"1d3d81e5-591a-469d-9851-86271162d455\") " pod="openstack/ovn-controller-kkdw8" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.283908 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1d3d81e5-591a-469d-9851-86271162d455-var-run\") pod \"ovn-controller-kkdw8\" (UID: \"1d3d81e5-591a-469d-9851-86271162d455\") " pod="openstack/ovn-controller-kkdw8" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.283928 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/e358b742-108b-4f60-989f-2e8eddce487e-var-lib\") pod \"ovn-controller-ovs-bkkst\" (UID: \"e358b742-108b-4f60-989f-2e8eddce487e\") " pod="openstack/ovn-controller-ovs-bkkst" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.283961 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d3d81e5-591a-469d-9851-86271162d455-ovn-controller-tls-certs\") pod \"ovn-controller-kkdw8\" (UID: \"1d3d81e5-591a-469d-9851-86271162d455\") " pod="openstack/ovn-controller-kkdw8" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.283985 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4n84\" (UniqueName: \"kubernetes.io/projected/e358b742-108b-4f60-989f-2e8eddce487e-kube-api-access-m4n84\") pod \"ovn-controller-ovs-bkkst\" (UID: \"e358b742-108b-4f60-989f-2e8eddce487e\") " pod="openstack/ovn-controller-ovs-bkkst" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.284011 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e358b742-108b-4f60-989f-2e8eddce487e-scripts\") pod \"ovn-controller-ovs-bkkst\" (UID: \"e358b742-108b-4f60-989f-2e8eddce487e\") " pod="openstack/ovn-controller-ovs-bkkst" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.284038 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x245m\" (UniqueName: \"kubernetes.io/projected/1d3d81e5-591a-469d-9851-86271162d455-kube-api-access-x245m\") pod \"ovn-controller-kkdw8\" (UID: \"1d3d81e5-591a-469d-9851-86271162d455\") " pod="openstack/ovn-controller-kkdw8" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.284058 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1d3d81e5-591a-469d-9851-86271162d455-var-run-ovn\") pod \"ovn-controller-kkdw8\" (UID: \"1d3d81e5-591a-469d-9851-86271162d455\") " pod="openstack/ovn-controller-kkdw8" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.284086 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1d3d81e5-591a-469d-9851-86271162d455-scripts\") pod \"ovn-controller-kkdw8\" (UID: \"1d3d81e5-591a-469d-9851-86271162d455\") " pod="openstack/ovn-controller-kkdw8" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.284514 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1d3d81e5-591a-469d-9851-86271162d455-var-log-ovn\") pod \"ovn-controller-kkdw8\" (UID: \"1d3d81e5-591a-469d-9851-86271162d455\") " pod="openstack/ovn-controller-kkdw8" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.284647 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1d3d81e5-591a-469d-9851-86271162d455-var-run\") pod \"ovn-controller-kkdw8\" (UID: \"1d3d81e5-591a-469d-9851-86271162d455\") " pod="openstack/ovn-controller-kkdw8" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.284706 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1d3d81e5-591a-469d-9851-86271162d455-var-run-ovn\") pod \"ovn-controller-kkdw8\" (UID: \"1d3d81e5-591a-469d-9851-86271162d455\") " pod="openstack/ovn-controller-kkdw8" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.286429 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1d3d81e5-591a-469d-9851-86271162d455-scripts\") pod \"ovn-controller-kkdw8\" (UID: \"1d3d81e5-591a-469d-9851-86271162d455\") " pod="openstack/ovn-controller-kkdw8" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.290781 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d3d81e5-591a-469d-9851-86271162d455-combined-ca-bundle\") pod \"ovn-controller-kkdw8\" (UID: \"1d3d81e5-591a-469d-9851-86271162d455\") " pod="openstack/ovn-controller-kkdw8" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.293531 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d3d81e5-591a-469d-9851-86271162d455-ovn-controller-tls-certs\") pod \"ovn-controller-kkdw8\" (UID: \"1d3d81e5-591a-469d-9851-86271162d455\") " pod="openstack/ovn-controller-kkdw8" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.304363 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x245m\" (UniqueName: \"kubernetes.io/projected/1d3d81e5-591a-469d-9851-86271162d455-kube-api-access-x245m\") pod \"ovn-controller-kkdw8\" (UID: \"1d3d81e5-591a-469d-9851-86271162d455\") " pod="openstack/ovn-controller-kkdw8" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.386296 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e358b742-108b-4f60-989f-2e8eddce487e-var-run\") pod \"ovn-controller-ovs-bkkst\" (UID: \"e358b742-108b-4f60-989f-2e8eddce487e\") " pod="openstack/ovn-controller-ovs-bkkst" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.386385 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/e358b742-108b-4f60-989f-2e8eddce487e-var-log\") pod \"ovn-controller-ovs-bkkst\" (UID: \"e358b742-108b-4f60-989f-2e8eddce487e\") " pod="openstack/ovn-controller-ovs-bkkst" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.386449 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/e358b742-108b-4f60-989f-2e8eddce487e-var-lib\") pod \"ovn-controller-ovs-bkkst\" (UID: \"e358b742-108b-4f60-989f-2e8eddce487e\") " pod="openstack/ovn-controller-ovs-bkkst" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.386497 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4n84\" (UniqueName: \"kubernetes.io/projected/e358b742-108b-4f60-989f-2e8eddce487e-kube-api-access-m4n84\") pod \"ovn-controller-ovs-bkkst\" (UID: \"e358b742-108b-4f60-989f-2e8eddce487e\") " pod="openstack/ovn-controller-ovs-bkkst" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.386544 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e358b742-108b-4f60-989f-2e8eddce487e-scripts\") pod \"ovn-controller-ovs-bkkst\" (UID: \"e358b742-108b-4f60-989f-2e8eddce487e\") " pod="openstack/ovn-controller-ovs-bkkst" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.386539 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e358b742-108b-4f60-989f-2e8eddce487e-var-run\") pod \"ovn-controller-ovs-bkkst\" (UID: \"e358b742-108b-4f60-989f-2e8eddce487e\") " pod="openstack/ovn-controller-ovs-bkkst" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.386706 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/e358b742-108b-4f60-989f-2e8eddce487e-etc-ovs\") pod \"ovn-controller-ovs-bkkst\" (UID: \"e358b742-108b-4f60-989f-2e8eddce487e\") " pod="openstack/ovn-controller-ovs-bkkst" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.386847 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/e358b742-108b-4f60-989f-2e8eddce487e-var-log\") pod \"ovn-controller-ovs-bkkst\" (UID: \"e358b742-108b-4f60-989f-2e8eddce487e\") " pod="openstack/ovn-controller-ovs-bkkst" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.387142 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/e358b742-108b-4f60-989f-2e8eddce487e-etc-ovs\") pod \"ovn-controller-ovs-bkkst\" (UID: \"e358b742-108b-4f60-989f-2e8eddce487e\") " pod="openstack/ovn-controller-ovs-bkkst" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.387208 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/e358b742-108b-4f60-989f-2e8eddce487e-var-lib\") pod \"ovn-controller-ovs-bkkst\" (UID: \"e358b742-108b-4f60-989f-2e8eddce487e\") " pod="openstack/ovn-controller-ovs-bkkst" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.389968 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e358b742-108b-4f60-989f-2e8eddce487e-scripts\") pod \"ovn-controller-ovs-bkkst\" (UID: \"e358b742-108b-4f60-989f-2e8eddce487e\") " pod="openstack/ovn-controller-ovs-bkkst" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.404525 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4n84\" (UniqueName: \"kubernetes.io/projected/e358b742-108b-4f60-989f-2e8eddce487e-kube-api-access-m4n84\") pod \"ovn-controller-ovs-bkkst\" (UID: \"e358b742-108b-4f60-989f-2e8eddce487e\") " pod="openstack/ovn-controller-ovs-bkkst" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.478355 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-kkdw8" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.537409 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.539325 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.542117 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-pmkn4" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.542839 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.543115 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.544528 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.545454 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.562830 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.577497 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-bkkst" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.590925 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"c79e4fc0-8786-40a6-a775-29a4767d072e\") " pod="openstack/ovsdbserver-nb-0" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.591034 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c79e4fc0-8786-40a6-a775-29a4767d072e-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"c79e4fc0-8786-40a6-a775-29a4767d072e\") " pod="openstack/ovsdbserver-nb-0" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.591434 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c79e4fc0-8786-40a6-a775-29a4767d072e-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c79e4fc0-8786-40a6-a775-29a4767d072e\") " pod="openstack/ovsdbserver-nb-0" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.591523 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c79e4fc0-8786-40a6-a775-29a4767d072e-config\") pod \"ovsdbserver-nb-0\" (UID: \"c79e4fc0-8786-40a6-a775-29a4767d072e\") " pod="openstack/ovsdbserver-nb-0" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.591567 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c79e4fc0-8786-40a6-a775-29a4767d072e-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"c79e4fc0-8786-40a6-a775-29a4767d072e\") " pod="openstack/ovsdbserver-nb-0" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.591603 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c79e4fc0-8786-40a6-a775-29a4767d072e-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"c79e4fc0-8786-40a6-a775-29a4767d072e\") " pod="openstack/ovsdbserver-nb-0" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.591646 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4v6w\" (UniqueName: \"kubernetes.io/projected/c79e4fc0-8786-40a6-a775-29a4767d072e-kube-api-access-w4v6w\") pod \"ovsdbserver-nb-0\" (UID: \"c79e4fc0-8786-40a6-a775-29a4767d072e\") " pod="openstack/ovsdbserver-nb-0" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.591727 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c79e4fc0-8786-40a6-a775-29a4767d072e-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c79e4fc0-8786-40a6-a775-29a4767d072e\") " pod="openstack/ovsdbserver-nb-0" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.693402 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"c79e4fc0-8786-40a6-a775-29a4767d072e\") " pod="openstack/ovsdbserver-nb-0" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.693469 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c79e4fc0-8786-40a6-a775-29a4767d072e-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"c79e4fc0-8786-40a6-a775-29a4767d072e\") " pod="openstack/ovsdbserver-nb-0" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.693500 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c79e4fc0-8786-40a6-a775-29a4767d072e-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c79e4fc0-8786-40a6-a775-29a4767d072e\") " pod="openstack/ovsdbserver-nb-0" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.693547 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c79e4fc0-8786-40a6-a775-29a4767d072e-config\") pod \"ovsdbserver-nb-0\" (UID: \"c79e4fc0-8786-40a6-a775-29a4767d072e\") " pod="openstack/ovsdbserver-nb-0" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.693566 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c79e4fc0-8786-40a6-a775-29a4767d072e-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"c79e4fc0-8786-40a6-a775-29a4767d072e\") " pod="openstack/ovsdbserver-nb-0" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.693587 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c79e4fc0-8786-40a6-a775-29a4767d072e-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"c79e4fc0-8786-40a6-a775-29a4767d072e\") " pod="openstack/ovsdbserver-nb-0" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.693610 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4v6w\" (UniqueName: \"kubernetes.io/projected/c79e4fc0-8786-40a6-a775-29a4767d072e-kube-api-access-w4v6w\") pod \"ovsdbserver-nb-0\" (UID: \"c79e4fc0-8786-40a6-a775-29a4767d072e\") " pod="openstack/ovsdbserver-nb-0" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.693684 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c79e4fc0-8786-40a6-a775-29a4767d072e-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c79e4fc0-8786-40a6-a775-29a4767d072e\") " pod="openstack/ovsdbserver-nb-0" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.695132 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c79e4fc0-8786-40a6-a775-29a4767d072e-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"c79e4fc0-8786-40a6-a775-29a4767d072e\") " pod="openstack/ovsdbserver-nb-0" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.695340 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c79e4fc0-8786-40a6-a775-29a4767d072e-config\") pod \"ovsdbserver-nb-0\" (UID: \"c79e4fc0-8786-40a6-a775-29a4767d072e\") " pod="openstack/ovsdbserver-nb-0" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.695463 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"c79e4fc0-8786-40a6-a775-29a4767d072e\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/ovsdbserver-nb-0" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.696347 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c79e4fc0-8786-40a6-a775-29a4767d072e-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"c79e4fc0-8786-40a6-a775-29a4767d072e\") " pod="openstack/ovsdbserver-nb-0" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.697528 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c79e4fc0-8786-40a6-a775-29a4767d072e-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c79e4fc0-8786-40a6-a775-29a4767d072e\") " pod="openstack/ovsdbserver-nb-0" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.699672 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c79e4fc0-8786-40a6-a775-29a4767d072e-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"c79e4fc0-8786-40a6-a775-29a4767d072e\") " pod="openstack/ovsdbserver-nb-0" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.700922 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c79e4fc0-8786-40a6-a775-29a4767d072e-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c79e4fc0-8786-40a6-a775-29a4767d072e\") " pod="openstack/ovsdbserver-nb-0" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.717156 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"c79e4fc0-8786-40a6-a775-29a4767d072e\") " pod="openstack/ovsdbserver-nb-0" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.728553 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4v6w\" (UniqueName: \"kubernetes.io/projected/c79e4fc0-8786-40a6-a775-29a4767d072e-kube-api-access-w4v6w\") pod \"ovsdbserver-nb-0\" (UID: \"c79e4fc0-8786-40a6-a775-29a4767d072e\") " pod="openstack/ovsdbserver-nb-0" Nov 29 04:53:21 crc kubenswrapper[4799]: I1129 04:53:21.857440 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 29 04:53:24 crc kubenswrapper[4799]: I1129 04:53:24.493494 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 29 04:53:24 crc kubenswrapper[4799]: I1129 04:53:24.495289 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 29 04:53:24 crc kubenswrapper[4799]: I1129 04:53:24.497698 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Nov 29 04:53:24 crc kubenswrapper[4799]: I1129 04:53:24.498119 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-sv5d4" Nov 29 04:53:24 crc kubenswrapper[4799]: I1129 04:53:24.498217 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Nov 29 04:53:24 crc kubenswrapper[4799]: I1129 04:53:24.498284 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Nov 29 04:53:24 crc kubenswrapper[4799]: I1129 04:53:24.513509 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 29 04:53:24 crc kubenswrapper[4799]: I1129 04:53:24.562035 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b276n\" (UniqueName: \"kubernetes.io/projected/5bb0e45f-f281-49ec-8a11-37e1ace553db-kube-api-access-b276n\") pod \"ovsdbserver-sb-0\" (UID: \"5bb0e45f-f281-49ec-8a11-37e1ace553db\") " pod="openstack/ovsdbserver-sb-0" Nov 29 04:53:24 crc kubenswrapper[4799]: I1129 04:53:24.562159 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bb0e45f-f281-49ec-8a11-37e1ace553db-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"5bb0e45f-f281-49ec-8a11-37e1ace553db\") " pod="openstack/ovsdbserver-sb-0" Nov 29 04:53:24 crc kubenswrapper[4799]: I1129 04:53:24.562189 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5bb0e45f-f281-49ec-8a11-37e1ace553db-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"5bb0e45f-f281-49ec-8a11-37e1ace553db\") " pod="openstack/ovsdbserver-sb-0" Nov 29 04:53:24 crc kubenswrapper[4799]: I1129 04:53:24.562270 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5bb0e45f-f281-49ec-8a11-37e1ace553db-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"5bb0e45f-f281-49ec-8a11-37e1ace553db\") " pod="openstack/ovsdbserver-sb-0" Nov 29 04:53:24 crc kubenswrapper[4799]: I1129 04:53:24.562311 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"5bb0e45f-f281-49ec-8a11-37e1ace553db\") " pod="openstack/ovsdbserver-sb-0" Nov 29 04:53:24 crc kubenswrapper[4799]: I1129 04:53:24.562333 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bb0e45f-f281-49ec-8a11-37e1ace553db-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"5bb0e45f-f281-49ec-8a11-37e1ace553db\") " pod="openstack/ovsdbserver-sb-0" Nov 29 04:53:24 crc kubenswrapper[4799]: I1129 04:53:24.562372 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bb0e45f-f281-49ec-8a11-37e1ace553db-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"5bb0e45f-f281-49ec-8a11-37e1ace553db\") " pod="openstack/ovsdbserver-sb-0" Nov 29 04:53:24 crc kubenswrapper[4799]: I1129 04:53:24.562397 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bb0e45f-f281-49ec-8a11-37e1ace553db-config\") pod \"ovsdbserver-sb-0\" (UID: \"5bb0e45f-f281-49ec-8a11-37e1ace553db\") " pod="openstack/ovsdbserver-sb-0" Nov 29 04:53:24 crc kubenswrapper[4799]: I1129 04:53:24.664502 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"5bb0e45f-f281-49ec-8a11-37e1ace553db\") " pod="openstack/ovsdbserver-sb-0" Nov 29 04:53:24 crc kubenswrapper[4799]: I1129 04:53:24.664551 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bb0e45f-f281-49ec-8a11-37e1ace553db-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"5bb0e45f-f281-49ec-8a11-37e1ace553db\") " pod="openstack/ovsdbserver-sb-0" Nov 29 04:53:24 crc kubenswrapper[4799]: I1129 04:53:24.665738 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bb0e45f-f281-49ec-8a11-37e1ace553db-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"5bb0e45f-f281-49ec-8a11-37e1ace553db\") " pod="openstack/ovsdbserver-sb-0" Nov 29 04:53:24 crc kubenswrapper[4799]: I1129 04:53:24.665773 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bb0e45f-f281-49ec-8a11-37e1ace553db-config\") pod \"ovsdbserver-sb-0\" (UID: \"5bb0e45f-f281-49ec-8a11-37e1ace553db\") " pod="openstack/ovsdbserver-sb-0" Nov 29 04:53:24 crc kubenswrapper[4799]: I1129 04:53:24.665834 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b276n\" (UniqueName: \"kubernetes.io/projected/5bb0e45f-f281-49ec-8a11-37e1ace553db-kube-api-access-b276n\") pod \"ovsdbserver-sb-0\" (UID: \"5bb0e45f-f281-49ec-8a11-37e1ace553db\") " pod="openstack/ovsdbserver-sb-0" Nov 29 04:53:24 crc kubenswrapper[4799]: I1129 04:53:24.665920 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bb0e45f-f281-49ec-8a11-37e1ace553db-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"5bb0e45f-f281-49ec-8a11-37e1ace553db\") " pod="openstack/ovsdbserver-sb-0" Nov 29 04:53:24 crc kubenswrapper[4799]: I1129 04:53:24.665951 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5bb0e45f-f281-49ec-8a11-37e1ace553db-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"5bb0e45f-f281-49ec-8a11-37e1ace553db\") " pod="openstack/ovsdbserver-sb-0" Nov 29 04:53:24 crc kubenswrapper[4799]: I1129 04:53:24.666087 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5bb0e45f-f281-49ec-8a11-37e1ace553db-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"5bb0e45f-f281-49ec-8a11-37e1ace553db\") " pod="openstack/ovsdbserver-sb-0" Nov 29 04:53:24 crc kubenswrapper[4799]: I1129 04:53:24.664923 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"5bb0e45f-f281-49ec-8a11-37e1ace553db\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/ovsdbserver-sb-0" Nov 29 04:53:24 crc kubenswrapper[4799]: I1129 04:53:24.666641 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5bb0e45f-f281-49ec-8a11-37e1ace553db-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"5bb0e45f-f281-49ec-8a11-37e1ace553db\") " pod="openstack/ovsdbserver-sb-0" Nov 29 04:53:24 crc kubenswrapper[4799]: I1129 04:53:24.667840 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5bb0e45f-f281-49ec-8a11-37e1ace553db-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"5bb0e45f-f281-49ec-8a11-37e1ace553db\") " pod="openstack/ovsdbserver-sb-0" Nov 29 04:53:24 crc kubenswrapper[4799]: I1129 04:53:24.668937 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bb0e45f-f281-49ec-8a11-37e1ace553db-config\") pod \"ovsdbserver-sb-0\" (UID: \"5bb0e45f-f281-49ec-8a11-37e1ace553db\") " pod="openstack/ovsdbserver-sb-0" Nov 29 04:53:24 crc kubenswrapper[4799]: I1129 04:53:24.675470 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bb0e45f-f281-49ec-8a11-37e1ace553db-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"5bb0e45f-f281-49ec-8a11-37e1ace553db\") " pod="openstack/ovsdbserver-sb-0" Nov 29 04:53:24 crc kubenswrapper[4799]: I1129 04:53:24.683637 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bb0e45f-f281-49ec-8a11-37e1ace553db-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"5bb0e45f-f281-49ec-8a11-37e1ace553db\") " pod="openstack/ovsdbserver-sb-0" Nov 29 04:53:24 crc kubenswrapper[4799]: I1129 04:53:24.684502 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bb0e45f-f281-49ec-8a11-37e1ace553db-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"5bb0e45f-f281-49ec-8a11-37e1ace553db\") " pod="openstack/ovsdbserver-sb-0" Nov 29 04:53:24 crc kubenswrapper[4799]: I1129 04:53:24.689177 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b276n\" (UniqueName: \"kubernetes.io/projected/5bb0e45f-f281-49ec-8a11-37e1ace553db-kube-api-access-b276n\") pod \"ovsdbserver-sb-0\" (UID: \"5bb0e45f-f281-49ec-8a11-37e1ace553db\") " pod="openstack/ovsdbserver-sb-0" Nov 29 04:53:24 crc kubenswrapper[4799]: I1129 04:53:24.691290 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"5bb0e45f-f281-49ec-8a11-37e1ace553db\") " pod="openstack/ovsdbserver-sb-0" Nov 29 04:53:24 crc kubenswrapper[4799]: I1129 04:53:24.820586 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 29 04:53:32 crc kubenswrapper[4799]: I1129 04:53:32.511188 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"bbc1f3a5-585d-4e6d-a87e-cf77a7b21147","Type":"ContainerStarted","Data":"24e3cca511a774dd98af6618e258a301d56c33bc1426594cd2dee1077fff2d15"} Nov 29 04:53:33 crc kubenswrapper[4799]: I1129 04:53:33.526670 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-knbck"] Nov 29 04:53:33 crc kubenswrapper[4799]: I1129 04:53:33.540709 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-knbck" Nov 29 04:53:33 crc kubenswrapper[4799]: I1129 04:53:33.542058 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-knbck"] Nov 29 04:53:33 crc kubenswrapper[4799]: I1129 04:53:33.642364 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f8a509f-be52-4f76-93e5-00f5dc59f7fb-catalog-content\") pod \"community-operators-knbck\" (UID: \"6f8a509f-be52-4f76-93e5-00f5dc59f7fb\") " pod="openshift-marketplace/community-operators-knbck" Nov 29 04:53:33 crc kubenswrapper[4799]: I1129 04:53:33.642434 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f8a509f-be52-4f76-93e5-00f5dc59f7fb-utilities\") pod \"community-operators-knbck\" (UID: \"6f8a509f-be52-4f76-93e5-00f5dc59f7fb\") " pod="openshift-marketplace/community-operators-knbck" Nov 29 04:53:33 crc kubenswrapper[4799]: I1129 04:53:33.642497 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdpx6\" (UniqueName: \"kubernetes.io/projected/6f8a509f-be52-4f76-93e5-00f5dc59f7fb-kube-api-access-xdpx6\") pod \"community-operators-knbck\" (UID: \"6f8a509f-be52-4f76-93e5-00f5dc59f7fb\") " pod="openshift-marketplace/community-operators-knbck" Nov 29 04:53:33 crc kubenswrapper[4799]: I1129 04:53:33.744339 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdpx6\" (UniqueName: \"kubernetes.io/projected/6f8a509f-be52-4f76-93e5-00f5dc59f7fb-kube-api-access-xdpx6\") pod \"community-operators-knbck\" (UID: \"6f8a509f-be52-4f76-93e5-00f5dc59f7fb\") " pod="openshift-marketplace/community-operators-knbck" Nov 29 04:53:33 crc kubenswrapper[4799]: I1129 04:53:33.744454 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f8a509f-be52-4f76-93e5-00f5dc59f7fb-catalog-content\") pod \"community-operators-knbck\" (UID: \"6f8a509f-be52-4f76-93e5-00f5dc59f7fb\") " pod="openshift-marketplace/community-operators-knbck" Nov 29 04:53:33 crc kubenswrapper[4799]: I1129 04:53:33.744486 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f8a509f-be52-4f76-93e5-00f5dc59f7fb-utilities\") pod \"community-operators-knbck\" (UID: \"6f8a509f-be52-4f76-93e5-00f5dc59f7fb\") " pod="openshift-marketplace/community-operators-knbck" Nov 29 04:53:33 crc kubenswrapper[4799]: I1129 04:53:33.745038 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f8a509f-be52-4f76-93e5-00f5dc59f7fb-utilities\") pod \"community-operators-knbck\" (UID: \"6f8a509f-be52-4f76-93e5-00f5dc59f7fb\") " pod="openshift-marketplace/community-operators-knbck" Nov 29 04:53:33 crc kubenswrapper[4799]: I1129 04:53:33.746577 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f8a509f-be52-4f76-93e5-00f5dc59f7fb-catalog-content\") pod \"community-operators-knbck\" (UID: \"6f8a509f-be52-4f76-93e5-00f5dc59f7fb\") " pod="openshift-marketplace/community-operators-knbck" Nov 29 04:53:33 crc kubenswrapper[4799]: I1129 04:53:33.775414 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdpx6\" (UniqueName: \"kubernetes.io/projected/6f8a509f-be52-4f76-93e5-00f5dc59f7fb-kube-api-access-xdpx6\") pod \"community-operators-knbck\" (UID: \"6f8a509f-be52-4f76-93e5-00f5dc59f7fb\") " pod="openshift-marketplace/community-operators-knbck" Nov 29 04:53:33 crc kubenswrapper[4799]: I1129 04:53:33.875620 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-knbck" Nov 29 04:53:41 crc kubenswrapper[4799]: E1129 04:53:41.524990 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Nov 29 04:53:41 crc kubenswrapper[4799]: E1129 04:53:41.525843 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-z859g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(81726039-72e9-455a-b4fb-e16184acdc47): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 29 04:53:41 crc kubenswrapper[4799]: E1129 04:53:41.527764 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="81726039-72e9-455a-b4fb-e16184acdc47" Nov 29 04:53:41 crc kubenswrapper[4799]: E1129 04:53:41.542396 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Nov 29 04:53:41 crc kubenswrapper[4799]: E1129 04:53:41.542649 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-k94v6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(1788e4d3-8c2a-4d06-b746-c0fb12514148): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 29 04:53:41 crc kubenswrapper[4799]: E1129 04:53:41.543928 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="1788e4d3-8c2a-4d06-b746-c0fb12514148" Nov 29 04:53:41 crc kubenswrapper[4799]: E1129 04:53:41.590971 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-server-0" podUID="1788e4d3-8c2a-4d06-b746-c0fb12514148" Nov 29 04:53:41 crc kubenswrapper[4799]: E1129 04:53:41.591063 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="81726039-72e9-455a-b4fb-e16184acdc47" Nov 29 04:53:42 crc kubenswrapper[4799]: I1129 04:53:42.127926 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 29 04:53:44 crc kubenswrapper[4799]: W1129 04:53:44.885232 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc79e4fc0_8786_40a6_a775_29a4767d072e.slice/crio-54b94d01a8e7838350c01c0dd34bd99cd0acd0e7c6a47522d760e238f04615c5 WatchSource:0}: Error finding container 54b94d01a8e7838350c01c0dd34bd99cd0acd0e7c6a47522d760e238f04615c5: Status 404 returned error can't find the container with id 54b94d01a8e7838350c01c0dd34bd99cd0acd0e7c6a47522d760e238f04615c5 Nov 29 04:53:44 crc kubenswrapper[4799]: E1129 04:53:44.886720 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 29 04:53:44 crc kubenswrapper[4799]: E1129 04:53:44.886970 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ccqck,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-9q9hd_openstack(da2af6d0-1824-4820-84eb-0c491cb777a9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 29 04:53:44 crc kubenswrapper[4799]: E1129 04:53:44.888120 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-9q9hd" podUID="da2af6d0-1824-4820-84eb-0c491cb777a9" Nov 29 04:53:44 crc kubenswrapper[4799]: E1129 04:53:44.911251 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 29 04:53:44 crc kubenswrapper[4799]: E1129 04:53:44.912381 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4pjhh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-vv6cv_openstack(a0f5fc9f-1ab2-42dc-9061-da1ae18610f1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 29 04:53:44 crc kubenswrapper[4799]: E1129 04:53:44.913684 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-vv6cv" podUID="a0f5fc9f-1ab2-42dc-9061-da1ae18610f1" Nov 29 04:53:44 crc kubenswrapper[4799]: E1129 04:53:44.929161 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 29 04:53:44 crc kubenswrapper[4799]: E1129 04:53:44.929401 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-crcdk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-ccgqx_openstack(f3627f09-e94d-4a00-94a2-7dcf7dfe7800): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 29 04:53:44 crc kubenswrapper[4799]: E1129 04:53:44.931440 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-ccgqx" podUID="f3627f09-e94d-4a00-94a2-7dcf7dfe7800" Nov 29 04:53:44 crc kubenswrapper[4799]: E1129 04:53:44.953625 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 29 04:53:44 crc kubenswrapper[4799]: E1129 04:53:44.953904 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4g6rd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-psbc4_openstack(99c095ca-db71-477a-9c51-3df3d2838144): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 29 04:53:44 crc kubenswrapper[4799]: E1129 04:53:44.955615 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-psbc4" podUID="99c095ca-db71-477a-9c51-3df3d2838144" Nov 29 04:53:45 crc kubenswrapper[4799]: I1129 04:53:45.627721 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"bbc1f3a5-585d-4e6d-a87e-cf77a7b21147","Type":"ContainerStarted","Data":"9b7f3479dd8c30364da8f14502f444499ff7ccdb8237d52a60b8f2d521e8e258"} Nov 29 04:53:45 crc kubenswrapper[4799]: I1129 04:53:45.630468 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"c79e4fc0-8786-40a6-a775-29a4767d072e","Type":"ContainerStarted","Data":"54b94d01a8e7838350c01c0dd34bd99cd0acd0e7c6a47522d760e238f04615c5"} Nov 29 04:53:45 crc kubenswrapper[4799]: E1129 04:53:45.633464 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-ccgqx" podUID="f3627f09-e94d-4a00-94a2-7dcf7dfe7800" Nov 29 04:53:45 crc kubenswrapper[4799]: E1129 04:53:45.637444 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-9q9hd" podUID="da2af6d0-1824-4820-84eb-0c491cb777a9" Nov 29 04:53:45 crc kubenswrapper[4799]: I1129 04:53:45.649054 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-knbck"] Nov 29 04:53:45 crc kubenswrapper[4799]: I1129 04:53:45.746332 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 29 04:53:45 crc kubenswrapper[4799]: I1129 04:53:45.757532 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-kkdw8"] Nov 29 04:53:45 crc kubenswrapper[4799]: I1129 04:53:45.779327 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 29 04:53:45 crc kubenswrapper[4799]: W1129 04:53:45.812943 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb6f5ef6d_bec8_4b82_8ebc_53952954134c.slice/crio-62dd8fa88cf7c072509297404067b2249255b6d9606087fd01336420ee2f8bf4 WatchSource:0}: Error finding container 62dd8fa88cf7c072509297404067b2249255b6d9606087fd01336420ee2f8bf4: Status 404 returned error can't find the container with id 62dd8fa88cf7c072509297404067b2249255b6d9606087fd01336420ee2f8bf4 Nov 29 04:53:45 crc kubenswrapper[4799]: I1129 04:53:45.851184 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 29 04:53:45 crc kubenswrapper[4799]: W1129 04:53:45.864696 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod80a2712a_3766_49b1_bab4_a806bb44e21f.slice/crio-2eb02c5a5b0ad0d1dc3d1e4d223c8ebb4b48b957d13ef9aa3d784da5967431b0 WatchSource:0}: Error finding container 2eb02c5a5b0ad0d1dc3d1e4d223c8ebb4b48b957d13ef9aa3d784da5967431b0: Status 404 returned error can't find the container with id 2eb02c5a5b0ad0d1dc3d1e4d223c8ebb4b48b957d13ef9aa3d784da5967431b0 Nov 29 04:53:45 crc kubenswrapper[4799]: I1129 04:53:45.876466 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7gtsn"] Nov 29 04:53:45 crc kubenswrapper[4799]: I1129 04:53:45.890775 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 29 04:53:45 crc kubenswrapper[4799]: I1129 04:53:45.899222 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-bkkst"] Nov 29 04:53:45 crc kubenswrapper[4799]: W1129 04:53:45.908426 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode358b742_108b_4f60_989f_2e8eddce487e.slice/crio-84b0e47abe9039d1123861d56b4be8ada79d2d0e19877dc8ed9d547632850300 WatchSource:0}: Error finding container 84b0e47abe9039d1123861d56b4be8ada79d2d0e19877dc8ed9d547632850300: Status 404 returned error can't find the container with id 84b0e47abe9039d1123861d56b4be8ada79d2d0e19877dc8ed9d547632850300 Nov 29 04:53:46 crc kubenswrapper[4799]: I1129 04:53:46.286284 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-psbc4" Nov 29 04:53:46 crc kubenswrapper[4799]: I1129 04:53:46.298021 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-vv6cv" Nov 29 04:53:46 crc kubenswrapper[4799]: I1129 04:53:46.366464 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99c095ca-db71-477a-9c51-3df3d2838144-config\") pod \"99c095ca-db71-477a-9c51-3df3d2838144\" (UID: \"99c095ca-db71-477a-9c51-3df3d2838144\") " Nov 29 04:53:46 crc kubenswrapper[4799]: I1129 04:53:46.366541 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4g6rd\" (UniqueName: \"kubernetes.io/projected/99c095ca-db71-477a-9c51-3df3d2838144-kube-api-access-4g6rd\") pod \"99c095ca-db71-477a-9c51-3df3d2838144\" (UID: \"99c095ca-db71-477a-9c51-3df3d2838144\") " Nov 29 04:53:46 crc kubenswrapper[4799]: I1129 04:53:46.367498 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99c095ca-db71-477a-9c51-3df3d2838144-config" (OuterVolumeSpecName: "config") pod "99c095ca-db71-477a-9c51-3df3d2838144" (UID: "99c095ca-db71-477a-9c51-3df3d2838144"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:53:46 crc kubenswrapper[4799]: I1129 04:53:46.388165 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99c095ca-db71-477a-9c51-3df3d2838144-kube-api-access-4g6rd" (OuterVolumeSpecName: "kube-api-access-4g6rd") pod "99c095ca-db71-477a-9c51-3df3d2838144" (UID: "99c095ca-db71-477a-9c51-3df3d2838144"). InnerVolumeSpecName "kube-api-access-4g6rd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:53:46 crc kubenswrapper[4799]: I1129 04:53:46.467938 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a0f5fc9f-1ab2-42dc-9061-da1ae18610f1-dns-svc\") pod \"a0f5fc9f-1ab2-42dc-9061-da1ae18610f1\" (UID: \"a0f5fc9f-1ab2-42dc-9061-da1ae18610f1\") " Nov 29 04:53:46 crc kubenswrapper[4799]: I1129 04:53:46.468045 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0f5fc9f-1ab2-42dc-9061-da1ae18610f1-config\") pod \"a0f5fc9f-1ab2-42dc-9061-da1ae18610f1\" (UID: \"a0f5fc9f-1ab2-42dc-9061-da1ae18610f1\") " Nov 29 04:53:46 crc kubenswrapper[4799]: I1129 04:53:46.468091 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4pjhh\" (UniqueName: \"kubernetes.io/projected/a0f5fc9f-1ab2-42dc-9061-da1ae18610f1-kube-api-access-4pjhh\") pod \"a0f5fc9f-1ab2-42dc-9061-da1ae18610f1\" (UID: \"a0f5fc9f-1ab2-42dc-9061-da1ae18610f1\") " Nov 29 04:53:46 crc kubenswrapper[4799]: I1129 04:53:46.468434 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99c095ca-db71-477a-9c51-3df3d2838144-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:53:46 crc kubenswrapper[4799]: I1129 04:53:46.468504 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4g6rd\" (UniqueName: \"kubernetes.io/projected/99c095ca-db71-477a-9c51-3df3d2838144-kube-api-access-4g6rd\") on node \"crc\" DevicePath \"\"" Nov 29 04:53:46 crc kubenswrapper[4799]: I1129 04:53:46.468610 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a0f5fc9f-1ab2-42dc-9061-da1ae18610f1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a0f5fc9f-1ab2-42dc-9061-da1ae18610f1" (UID: "a0f5fc9f-1ab2-42dc-9061-da1ae18610f1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:53:46 crc kubenswrapper[4799]: I1129 04:53:46.469228 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a0f5fc9f-1ab2-42dc-9061-da1ae18610f1-config" (OuterVolumeSpecName: "config") pod "a0f5fc9f-1ab2-42dc-9061-da1ae18610f1" (UID: "a0f5fc9f-1ab2-42dc-9061-da1ae18610f1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:53:46 crc kubenswrapper[4799]: I1129 04:53:46.471052 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0f5fc9f-1ab2-42dc-9061-da1ae18610f1-kube-api-access-4pjhh" (OuterVolumeSpecName: "kube-api-access-4pjhh") pod "a0f5fc9f-1ab2-42dc-9061-da1ae18610f1" (UID: "a0f5fc9f-1ab2-42dc-9061-da1ae18610f1"). InnerVolumeSpecName "kube-api-access-4pjhh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:53:46 crc kubenswrapper[4799]: I1129 04:53:46.574377 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a0f5fc9f-1ab2-42dc-9061-da1ae18610f1-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 29 04:53:46 crc kubenswrapper[4799]: I1129 04:53:46.574406 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0f5fc9f-1ab2-42dc-9061-da1ae18610f1-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:53:46 crc kubenswrapper[4799]: I1129 04:53:46.574417 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4pjhh\" (UniqueName: \"kubernetes.io/projected/a0f5fc9f-1ab2-42dc-9061-da1ae18610f1-kube-api-access-4pjhh\") on node \"crc\" DevicePath \"\"" Nov 29 04:53:46 crc kubenswrapper[4799]: I1129 04:53:46.638328 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-psbc4" event={"ID":"99c095ca-db71-477a-9c51-3df3d2838144","Type":"ContainerDied","Data":"2b37b1b7b391fb353165defa9aefe2718ce43a42c7382dfbc9858a3f47a69028"} Nov 29 04:53:46 crc kubenswrapper[4799]: I1129 04:53:46.638598 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-psbc4" Nov 29 04:53:46 crc kubenswrapper[4799]: I1129 04:53:46.737155 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"b6f5ef6d-bec8-4b82-8ebc-53952954134c","Type":"ContainerStarted","Data":"62dd8fa88cf7c072509297404067b2249255b6d9606087fd01336420ee2f8bf4"} Nov 29 04:53:46 crc kubenswrapper[4799]: I1129 04:53:46.784159 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"5bb0e45f-f281-49ec-8a11-37e1ace553db","Type":"ContainerStarted","Data":"4cbab332563df2cf5dea1dabd524513b39895ec68ffce78deb2297f46be9e41b"} Nov 29 04:53:46 crc kubenswrapper[4799]: I1129 04:53:46.852577 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-psbc4"] Nov 29 04:53:46 crc kubenswrapper[4799]: I1129 04:53:46.855058 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"735902f6-652e-4c85-818d-10f7d9529788","Type":"ContainerStarted","Data":"9cb9e9319ea2df9e2116660eac0c9a39a51288b0f4a9d6e63de601cf9134e055"} Nov 29 04:53:46 crc kubenswrapper[4799]: I1129 04:53:46.855104 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"735902f6-652e-4c85-818d-10f7d9529788","Type":"ContainerStarted","Data":"965d1436634c5f29b67fcba72036d81855f8e74a215835beddb0ee06aae211e3"} Nov 29 04:53:46 crc kubenswrapper[4799]: I1129 04:53:46.860446 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-psbc4"] Nov 29 04:53:46 crc kubenswrapper[4799]: I1129 04:53:46.867932 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-bkkst" event={"ID":"e358b742-108b-4f60-989f-2e8eddce487e","Type":"ContainerStarted","Data":"84b0e47abe9039d1123861d56b4be8ada79d2d0e19877dc8ed9d547632850300"} Nov 29 04:53:46 crc kubenswrapper[4799]: I1129 04:53:46.869859 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-vv6cv" event={"ID":"a0f5fc9f-1ab2-42dc-9061-da1ae18610f1","Type":"ContainerDied","Data":"486a647bff80499971d4e626bb10a5d919d6a4f6541a0060539c166ec316d912"} Nov 29 04:53:46 crc kubenswrapper[4799]: I1129 04:53:46.869956 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-vv6cv" Nov 29 04:53:46 crc kubenswrapper[4799]: I1129 04:53:46.873594 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"9c406f33-9e26-4226-b6dd-c7449fa96b9f","Type":"ContainerStarted","Data":"8d2c691761531ac2ef5a5c98cb36d8c73b2daef63c1d86a04566795ca6be7493"} Nov 29 04:53:46 crc kubenswrapper[4799]: I1129 04:53:46.875013 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-kkdw8" event={"ID":"1d3d81e5-591a-469d-9851-86271162d455","Type":"ContainerStarted","Data":"2a52116bfd6345e6584f5b4a17f1675cb295c58fc943e2c14803ba589d1d9dda"} Nov 29 04:53:46 crc kubenswrapper[4799]: I1129 04:53:46.878284 4799 generic.go:334] "Generic (PLEG): container finished" podID="80a2712a-3766-49b1-bab4-a806bb44e21f" containerID="148179253a550e8c180d17b00014302c25693e63366864ce710dc422e93e0425" exitCode=0 Nov 29 04:53:46 crc kubenswrapper[4799]: I1129 04:53:46.878326 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7gtsn" event={"ID":"80a2712a-3766-49b1-bab4-a806bb44e21f","Type":"ContainerDied","Data":"148179253a550e8c180d17b00014302c25693e63366864ce710dc422e93e0425"} Nov 29 04:53:46 crc kubenswrapper[4799]: I1129 04:53:46.878347 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7gtsn" event={"ID":"80a2712a-3766-49b1-bab4-a806bb44e21f","Type":"ContainerStarted","Data":"2eb02c5a5b0ad0d1dc3d1e4d223c8ebb4b48b957d13ef9aa3d784da5967431b0"} Nov 29 04:53:46 crc kubenswrapper[4799]: I1129 04:53:46.895466 4799 generic.go:334] "Generic (PLEG): container finished" podID="6f8a509f-be52-4f76-93e5-00f5dc59f7fb" containerID="120d5265ef50f957c208ec2016b4dc73af0337344e39abc882539af438491617" exitCode=0 Nov 29 04:53:46 crc kubenswrapper[4799]: I1129 04:53:46.895596 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-knbck" event={"ID":"6f8a509f-be52-4f76-93e5-00f5dc59f7fb","Type":"ContainerDied","Data":"120d5265ef50f957c208ec2016b4dc73af0337344e39abc882539af438491617"} Nov 29 04:53:46 crc kubenswrapper[4799]: I1129 04:53:46.895659 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-knbck" event={"ID":"6f8a509f-be52-4f76-93e5-00f5dc59f7fb","Type":"ContainerStarted","Data":"724a4426084693be9ded30332cfe8a5921be89cf719acb22d002be64f952fbff"} Nov 29 04:53:47 crc kubenswrapper[4799]: I1129 04:53:47.022107 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-vv6cv"] Nov 29 04:53:47 crc kubenswrapper[4799]: I1129 04:53:47.033104 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-vv6cv"] Nov 29 04:53:48 crc kubenswrapper[4799]: I1129 04:53:48.688523 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99c095ca-db71-477a-9c51-3df3d2838144" path="/var/lib/kubelet/pods/99c095ca-db71-477a-9c51-3df3d2838144/volumes" Nov 29 04:53:48 crc kubenswrapper[4799]: I1129 04:53:48.689442 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0f5fc9f-1ab2-42dc-9061-da1ae18610f1" path="/var/lib/kubelet/pods/a0f5fc9f-1ab2-42dc-9061-da1ae18610f1/volumes" Nov 29 04:53:49 crc kubenswrapper[4799]: I1129 04:53:49.924195 4799 generic.go:334] "Generic (PLEG): container finished" podID="bbc1f3a5-585d-4e6d-a87e-cf77a7b21147" containerID="9b7f3479dd8c30364da8f14502f444499ff7ccdb8237d52a60b8f2d521e8e258" exitCode=0 Nov 29 04:53:49 crc kubenswrapper[4799]: I1129 04:53:49.924252 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"bbc1f3a5-585d-4e6d-a87e-cf77a7b21147","Type":"ContainerDied","Data":"9b7f3479dd8c30364da8f14502f444499ff7ccdb8237d52a60b8f2d521e8e258"} Nov 29 04:53:50 crc kubenswrapper[4799]: I1129 04:53:50.935122 4799 generic.go:334] "Generic (PLEG): container finished" podID="735902f6-652e-4c85-818d-10f7d9529788" containerID="9cb9e9319ea2df9e2116660eac0c9a39a51288b0f4a9d6e63de601cf9134e055" exitCode=0 Nov 29 04:53:50 crc kubenswrapper[4799]: I1129 04:53:50.935212 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"735902f6-652e-4c85-818d-10f7d9529788","Type":"ContainerDied","Data":"9cb9e9319ea2df9e2116660eac0c9a39a51288b0f4a9d6e63de601cf9134e055"} Nov 29 04:53:54 crc kubenswrapper[4799]: I1129 04:53:54.935762 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jpq6g"] Nov 29 04:53:54 crc kubenswrapper[4799]: I1129 04:53:54.939011 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jpq6g" Nov 29 04:53:54 crc kubenswrapper[4799]: I1129 04:53:54.944411 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jpq6g"] Nov 29 04:53:54 crc kubenswrapper[4799]: I1129 04:53:54.978742 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89b0a912-606b-4d5d-bd40-09f89f2db649-utilities\") pod \"redhat-operators-jpq6g\" (UID: \"89b0a912-606b-4d5d-bd40-09f89f2db649\") " pod="openshift-marketplace/redhat-operators-jpq6g" Nov 29 04:53:54 crc kubenswrapper[4799]: I1129 04:53:54.979095 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfl99\" (UniqueName: \"kubernetes.io/projected/89b0a912-606b-4d5d-bd40-09f89f2db649-kube-api-access-jfl99\") pod \"redhat-operators-jpq6g\" (UID: \"89b0a912-606b-4d5d-bd40-09f89f2db649\") " pod="openshift-marketplace/redhat-operators-jpq6g" Nov 29 04:53:54 crc kubenswrapper[4799]: I1129 04:53:54.979232 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89b0a912-606b-4d5d-bd40-09f89f2db649-catalog-content\") pod \"redhat-operators-jpq6g\" (UID: \"89b0a912-606b-4d5d-bd40-09f89f2db649\") " pod="openshift-marketplace/redhat-operators-jpq6g" Nov 29 04:53:55 crc kubenswrapper[4799]: I1129 04:53:55.082338 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfl99\" (UniqueName: \"kubernetes.io/projected/89b0a912-606b-4d5d-bd40-09f89f2db649-kube-api-access-jfl99\") pod \"redhat-operators-jpq6g\" (UID: \"89b0a912-606b-4d5d-bd40-09f89f2db649\") " pod="openshift-marketplace/redhat-operators-jpq6g" Nov 29 04:53:55 crc kubenswrapper[4799]: I1129 04:53:55.082426 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89b0a912-606b-4d5d-bd40-09f89f2db649-catalog-content\") pod \"redhat-operators-jpq6g\" (UID: \"89b0a912-606b-4d5d-bd40-09f89f2db649\") " pod="openshift-marketplace/redhat-operators-jpq6g" Nov 29 04:53:55 crc kubenswrapper[4799]: I1129 04:53:55.082535 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89b0a912-606b-4d5d-bd40-09f89f2db649-utilities\") pod \"redhat-operators-jpq6g\" (UID: \"89b0a912-606b-4d5d-bd40-09f89f2db649\") " pod="openshift-marketplace/redhat-operators-jpq6g" Nov 29 04:53:55 crc kubenswrapper[4799]: I1129 04:53:55.083126 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89b0a912-606b-4d5d-bd40-09f89f2db649-utilities\") pod \"redhat-operators-jpq6g\" (UID: \"89b0a912-606b-4d5d-bd40-09f89f2db649\") " pod="openshift-marketplace/redhat-operators-jpq6g" Nov 29 04:53:55 crc kubenswrapper[4799]: I1129 04:53:55.083535 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89b0a912-606b-4d5d-bd40-09f89f2db649-catalog-content\") pod \"redhat-operators-jpq6g\" (UID: \"89b0a912-606b-4d5d-bd40-09f89f2db649\") " pod="openshift-marketplace/redhat-operators-jpq6g" Nov 29 04:53:55 crc kubenswrapper[4799]: I1129 04:53:55.105639 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfl99\" (UniqueName: \"kubernetes.io/projected/89b0a912-606b-4d5d-bd40-09f89f2db649-kube-api-access-jfl99\") pod \"redhat-operators-jpq6g\" (UID: \"89b0a912-606b-4d5d-bd40-09f89f2db649\") " pod="openshift-marketplace/redhat-operators-jpq6g" Nov 29 04:53:55 crc kubenswrapper[4799]: I1129 04:53:55.440749 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jpq6g" Nov 29 04:53:56 crc kubenswrapper[4799]: I1129 04:53:56.010408 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jpq6g"] Nov 29 04:53:56 crc kubenswrapper[4799]: I1129 04:53:56.023559 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"b6f5ef6d-bec8-4b82-8ebc-53952954134c","Type":"ContainerStarted","Data":"da3969a2bf92ad16ad96c3e0612ee755c95a4135a284ddeac8c6521ae1827ee1"} Nov 29 04:53:56 crc kubenswrapper[4799]: I1129 04:53:56.025172 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Nov 29 04:53:56 crc kubenswrapper[4799]: I1129 04:53:56.028267 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"5bb0e45f-f281-49ec-8a11-37e1ace553db","Type":"ContainerStarted","Data":"ff33969558720bbe6b4e808c73a3f5eb81a583fad93ddbf734ac492bd6a7d67f"} Nov 29 04:53:56 crc kubenswrapper[4799]: I1129 04:53:56.033860 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"bbc1f3a5-585d-4e6d-a87e-cf77a7b21147","Type":"ContainerStarted","Data":"adefb1090e70538b991f8319f2b3e903f2deff3e95b745cb910f88970a3dcadd"} Nov 29 04:53:56 crc kubenswrapper[4799]: I1129 04:53:56.044300 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-kkdw8" event={"ID":"1d3d81e5-591a-469d-9851-86271162d455","Type":"ContainerStarted","Data":"3c41b8d821dadca341e6fa9c219c4775bae4e925d34867920363f867ba7a4a23"} Nov 29 04:53:56 crc kubenswrapper[4799]: I1129 04:53:56.045242 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-kkdw8" Nov 29 04:53:56 crc kubenswrapper[4799]: I1129 04:53:56.046680 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=34.522947162 podStartE2EDuration="41.04665373s" podCreationTimestamp="2025-11-29 04:53:15 +0000 UTC" firstStartedPulling="2025-11-29 04:53:45.821518296 +0000 UTC m=+901.464448696" lastFinishedPulling="2025-11-29 04:53:52.345224864 +0000 UTC m=+907.988155264" observedRunningTime="2025-11-29 04:53:56.045554495 +0000 UTC m=+911.688484895" watchObservedRunningTime="2025-11-29 04:53:56.04665373 +0000 UTC m=+911.689584130" Nov 29 04:53:56 crc kubenswrapper[4799]: I1129 04:53:56.068024 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"735902f6-652e-4c85-818d-10f7d9529788","Type":"ContainerStarted","Data":"a09c16ed4e00d6ea88dbbd907a7a88ec727dd512799c3ed5ccfeada9c856e380"} Nov 29 04:53:56 crc kubenswrapper[4799]: I1129 04:53:56.071694 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"9c406f33-9e26-4226-b6dd-c7449fa96b9f","Type":"ContainerStarted","Data":"ab9bfe5165025b9af4a628e94c467832cbbce81cb0f59b89fbbe2bd8f224ef8d"} Nov 29 04:53:56 crc kubenswrapper[4799]: I1129 04:53:56.072372 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 29 04:53:56 crc kubenswrapper[4799]: I1129 04:53:56.080536 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"c79e4fc0-8786-40a6-a775-29a4767d072e","Type":"ContainerStarted","Data":"ac87bf68daa4c769ce468f180f3250910963449ec4b852b50c2903642f7e3b99"} Nov 29 04:53:56 crc kubenswrapper[4799]: I1129 04:53:56.091763 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-kkdw8" podStartSLOduration=26.921584299 podStartE2EDuration="35.091737605s" podCreationTimestamp="2025-11-29 04:53:21 +0000 UTC" firstStartedPulling="2025-11-29 04:53:45.789704665 +0000 UTC m=+901.432635065" lastFinishedPulling="2025-11-29 04:53:53.959857981 +0000 UTC m=+909.602788371" observedRunningTime="2025-11-29 04:53:56.072652294 +0000 UTC m=+911.715582694" watchObservedRunningTime="2025-11-29 04:53:56.091737605 +0000 UTC m=+911.734668005" Nov 29 04:53:56 crc kubenswrapper[4799]: I1129 04:53:56.093742 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-bkkst" event={"ID":"e358b742-108b-4f60-989f-2e8eddce487e","Type":"ContainerStarted","Data":"9ac874ad25326986ca9dcdb2483c67a86ad5bfaf7a0e090ed55ecdd88b60930e"} Nov 29 04:53:56 crc kubenswrapper[4799]: I1129 04:53:56.098455 4799 generic.go:334] "Generic (PLEG): container finished" podID="80a2712a-3766-49b1-bab4-a806bb44e21f" containerID="401ec315c842097520dc3a517a3d7f83b9ed67273f265d705da6318a2b3e540c" exitCode=0 Nov 29 04:53:56 crc kubenswrapper[4799]: I1129 04:53:56.098772 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7gtsn" event={"ID":"80a2712a-3766-49b1-bab4-a806bb44e21f","Type":"ContainerDied","Data":"401ec315c842097520dc3a517a3d7f83b9ed67273f265d705da6318a2b3e540c"} Nov 29 04:53:56 crc kubenswrapper[4799]: I1129 04:53:56.103166 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-knbck" event={"ID":"6f8a509f-be52-4f76-93e5-00f5dc59f7fb","Type":"ContainerStarted","Data":"8ad8330f282239422cacd78f34b97a74be86fb631e610218e468bf9191dcf024"} Nov 29 04:53:56 crc kubenswrapper[4799]: I1129 04:53:56.106268 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=31.345878404 podStartE2EDuration="44.106255038s" podCreationTimestamp="2025-11-29 04:53:12 +0000 UTC" firstStartedPulling="2025-11-29 04:53:32.213833848 +0000 UTC m=+887.856764258" lastFinishedPulling="2025-11-29 04:53:44.974210492 +0000 UTC m=+900.617140892" observedRunningTime="2025-11-29 04:53:56.10509262 +0000 UTC m=+911.748023020" watchObservedRunningTime="2025-11-29 04:53:56.106255038 +0000 UTC m=+911.749185438" Nov 29 04:53:56 crc kubenswrapper[4799]: I1129 04:53:56.137831 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=43.137782703 podStartE2EDuration="43.137782703s" podCreationTimestamp="2025-11-29 04:53:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:53:56.128727579 +0000 UTC m=+911.771657979" watchObservedRunningTime="2025-11-29 04:53:56.137782703 +0000 UTC m=+911.780713103" Nov 29 04:53:56 crc kubenswrapper[4799]: I1129 04:53:56.223208 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=31.054057168 podStartE2EDuration="40.22317689s" podCreationTimestamp="2025-11-29 04:53:16 +0000 UTC" firstStartedPulling="2025-11-29 04:53:45.806555893 +0000 UTC m=+901.449486283" lastFinishedPulling="2025-11-29 04:53:54.975675605 +0000 UTC m=+910.618606005" observedRunningTime="2025-11-29 04:53:56.211202537 +0000 UTC m=+911.854132937" watchObservedRunningTime="2025-11-29 04:53:56.22317689 +0000 UTC m=+911.866107290" Nov 29 04:53:56 crc kubenswrapper[4799]: E1129 04:53:56.474997 4799 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.83:56898->38.102.83.83:33865: write tcp 38.102.83.83:56898->38.102.83.83:33865: write: broken pipe Nov 29 04:53:57 crc kubenswrapper[4799]: I1129 04:53:57.119624 4799 generic.go:334] "Generic (PLEG): container finished" podID="e358b742-108b-4f60-989f-2e8eddce487e" containerID="9ac874ad25326986ca9dcdb2483c67a86ad5bfaf7a0e090ed55ecdd88b60930e" exitCode=0 Nov 29 04:53:57 crc kubenswrapper[4799]: I1129 04:53:57.119727 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-bkkst" event={"ID":"e358b742-108b-4f60-989f-2e8eddce487e","Type":"ContainerDied","Data":"9ac874ad25326986ca9dcdb2483c67a86ad5bfaf7a0e090ed55ecdd88b60930e"} Nov 29 04:53:57 crc kubenswrapper[4799]: I1129 04:53:57.138838 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7gtsn" event={"ID":"80a2712a-3766-49b1-bab4-a806bb44e21f","Type":"ContainerStarted","Data":"44dd2618c30517ff8ffaf8fc851a2767b7bf67324399a981f948ffecae837bcc"} Nov 29 04:53:57 crc kubenswrapper[4799]: I1129 04:53:57.144140 4799 generic.go:334] "Generic (PLEG): container finished" podID="6f8a509f-be52-4f76-93e5-00f5dc59f7fb" containerID="8ad8330f282239422cacd78f34b97a74be86fb631e610218e468bf9191dcf024" exitCode=0 Nov 29 04:53:57 crc kubenswrapper[4799]: I1129 04:53:57.144265 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-knbck" event={"ID":"6f8a509f-be52-4f76-93e5-00f5dc59f7fb","Type":"ContainerDied","Data":"8ad8330f282239422cacd78f34b97a74be86fb631e610218e468bf9191dcf024"} Nov 29 04:53:57 crc kubenswrapper[4799]: I1129 04:53:57.150075 4799 generic.go:334] "Generic (PLEG): container finished" podID="89b0a912-606b-4d5d-bd40-09f89f2db649" containerID="782739f273c3e17b3a2fe4efbe5a5517a28b476625a999543621bd4b130b3ffb" exitCode=0 Nov 29 04:53:57 crc kubenswrapper[4799]: I1129 04:53:57.150130 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jpq6g" event={"ID":"89b0a912-606b-4d5d-bd40-09f89f2db649","Type":"ContainerDied","Data":"782739f273c3e17b3a2fe4efbe5a5517a28b476625a999543621bd4b130b3ffb"} Nov 29 04:53:57 crc kubenswrapper[4799]: I1129 04:53:57.150161 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jpq6g" event={"ID":"89b0a912-606b-4d5d-bd40-09f89f2db649","Type":"ContainerStarted","Data":"3b1bbf0db10e104f84595b8b29c5cddadb967e17cfa5e7c40da5f7326e0850ce"} Nov 29 04:53:57 crc kubenswrapper[4799]: I1129 04:53:57.152493 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1788e4d3-8c2a-4d06-b746-c0fb12514148","Type":"ContainerStarted","Data":"c7803748c9deb308ab6a3661ce9b6ea289206d8f547aa0262d26c7a7a46628cc"} Nov 29 04:53:57 crc kubenswrapper[4799]: I1129 04:53:57.154465 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"81726039-72e9-455a-b4fb-e16184acdc47","Type":"ContainerStarted","Data":"a72d02dc2cf8aa7523907bde01f61568e9db9f50c267ea96039aa2db0a285b08"} Nov 29 04:53:57 crc kubenswrapper[4799]: I1129 04:53:57.168271 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7gtsn" podStartSLOduration=29.499352514999998 podStartE2EDuration="39.168247542s" podCreationTimestamp="2025-11-29 04:53:18 +0000 UTC" firstStartedPulling="2025-11-29 04:53:46.882188158 +0000 UTC m=+902.525118558" lastFinishedPulling="2025-11-29 04:53:56.551083185 +0000 UTC m=+912.194013585" observedRunningTime="2025-11-29 04:53:57.159868804 +0000 UTC m=+912.802799204" watchObservedRunningTime="2025-11-29 04:53:57.168247542 +0000 UTC m=+912.811177942" Nov 29 04:53:58 crc kubenswrapper[4799]: I1129 04:53:58.169446 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-bkkst" event={"ID":"e358b742-108b-4f60-989f-2e8eddce487e","Type":"ContainerStarted","Data":"5ba82ddea94756bb50265eedb667851d7b0e805888dae504b62ffdaceaf87659"} Nov 29 04:53:58 crc kubenswrapper[4799]: I1129 04:53:58.170058 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-bkkst" event={"ID":"e358b742-108b-4f60-989f-2e8eddce487e","Type":"ContainerStarted","Data":"2c81d213083bd4d1c7cabbb5624f5f5c7f21821fd47022d51df86a35ceda4238"} Nov 29 04:53:58 crc kubenswrapper[4799]: I1129 04:53:58.170078 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-bkkst" Nov 29 04:53:58 crc kubenswrapper[4799]: I1129 04:53:58.170090 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-bkkst" Nov 29 04:53:58 crc kubenswrapper[4799]: I1129 04:53:58.195944 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-knbck" event={"ID":"6f8a509f-be52-4f76-93e5-00f5dc59f7fb","Type":"ContainerStarted","Data":"cebadcc259c75c7a535aebe5e9c044d01c76e8f7e2ad239e287b3c3ab5f063f8"} Nov 29 04:53:58 crc kubenswrapper[4799]: I1129 04:53:58.199392 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jpq6g" event={"ID":"89b0a912-606b-4d5d-bd40-09f89f2db649","Type":"ContainerStarted","Data":"54931734410bb6d7974694eac8365871e27917bb9d9e0801647e259119412efe"} Nov 29 04:53:58 crc kubenswrapper[4799]: I1129 04:53:58.213283 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-bkkst" podStartSLOduration=30.695178831 podStartE2EDuration="37.213249935s" podCreationTimestamp="2025-11-29 04:53:21 +0000 UTC" firstStartedPulling="2025-11-29 04:53:45.917209946 +0000 UTC m=+901.560140346" lastFinishedPulling="2025-11-29 04:53:52.43528105 +0000 UTC m=+908.078211450" observedRunningTime="2025-11-29 04:53:58.210130781 +0000 UTC m=+913.853061181" watchObservedRunningTime="2025-11-29 04:53:58.213249935 +0000 UTC m=+913.856180335" Nov 29 04:53:58 crc kubenswrapper[4799]: I1129 04:53:58.264666 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-knbck" podStartSLOduration=14.444491212 podStartE2EDuration="25.264638629s" podCreationTimestamp="2025-11-29 04:53:33 +0000 UTC" firstStartedPulling="2025-11-29 04:53:46.898279969 +0000 UTC m=+902.541210369" lastFinishedPulling="2025-11-29 04:53:57.718427386 +0000 UTC m=+913.361357786" observedRunningTime="2025-11-29 04:53:58.25834525 +0000 UTC m=+913.901275650" watchObservedRunningTime="2025-11-29 04:53:58.264638629 +0000 UTC m=+913.907569029" Nov 29 04:53:59 crc kubenswrapper[4799]: I1129 04:53:59.222995 4799 generic.go:334] "Generic (PLEG): container finished" podID="89b0a912-606b-4d5d-bd40-09f89f2db649" containerID="54931734410bb6d7974694eac8365871e27917bb9d9e0801647e259119412efe" exitCode=0 Nov 29 04:53:59 crc kubenswrapper[4799]: I1129 04:53:59.225404 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jpq6g" event={"ID":"89b0a912-606b-4d5d-bd40-09f89f2db649","Type":"ContainerDied","Data":"54931734410bb6d7974694eac8365871e27917bb9d9e0801647e259119412efe"} Nov 29 04:53:59 crc kubenswrapper[4799]: I1129 04:53:59.323150 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7gtsn" Nov 29 04:53:59 crc kubenswrapper[4799]: I1129 04:53:59.323231 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7gtsn" Nov 29 04:53:59 crc kubenswrapper[4799]: I1129 04:53:59.370414 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7gtsn" Nov 29 04:54:00 crc kubenswrapper[4799]: I1129 04:54:00.384985 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Nov 29 04:54:03 crc kubenswrapper[4799]: I1129 04:54:03.814137 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Nov 29 04:54:03 crc kubenswrapper[4799]: I1129 04:54:03.815682 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Nov 29 04:54:03 crc kubenswrapper[4799]: I1129 04:54:03.876360 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-knbck" Nov 29 04:54:03 crc kubenswrapper[4799]: I1129 04:54:03.876488 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-knbck" Nov 29 04:54:03 crc kubenswrapper[4799]: I1129 04:54:03.925466 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-knbck" Nov 29 04:54:04 crc kubenswrapper[4799]: I1129 04:54:04.308159 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-knbck" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.132412 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-5szsr"] Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.133741 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-5szsr" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.138956 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.157874 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-5szsr"] Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.167137 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-knbck"] Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.253238 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.253413 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.292676 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/b0d261b9-8134-46ec-b510-63fce197ccff-ovn-rundir\") pod \"ovn-controller-metrics-5szsr\" (UID: \"b0d261b9-8134-46ec-b510-63fce197ccff\") " pod="openstack/ovn-controller-metrics-5szsr" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.292817 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/b0d261b9-8134-46ec-b510-63fce197ccff-ovs-rundir\") pod \"ovn-controller-metrics-5szsr\" (UID: \"b0d261b9-8134-46ec-b510-63fce197ccff\") " pod="openstack/ovn-controller-metrics-5szsr" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.292873 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0d261b9-8134-46ec-b510-63fce197ccff-config\") pod \"ovn-controller-metrics-5szsr\" (UID: \"b0d261b9-8134-46ec-b510-63fce197ccff\") " pod="openstack/ovn-controller-metrics-5szsr" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.292903 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0d261b9-8134-46ec-b510-63fce197ccff-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-5szsr\" (UID: \"b0d261b9-8134-46ec-b510-63fce197ccff\") " pod="openstack/ovn-controller-metrics-5szsr" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.292972 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0d261b9-8134-46ec-b510-63fce197ccff-combined-ca-bundle\") pod \"ovn-controller-metrics-5szsr\" (UID: \"b0d261b9-8134-46ec-b510-63fce197ccff\") " pod="openstack/ovn-controller-metrics-5szsr" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.293006 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j46xl\" (UniqueName: \"kubernetes.io/projected/b0d261b9-8134-46ec-b510-63fce197ccff-kube-api-access-j46xl\") pod \"ovn-controller-metrics-5szsr\" (UID: \"b0d261b9-8134-46ec-b510-63fce197ccff\") " pod="openstack/ovn-controller-metrics-5szsr" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.304330 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-446gh"] Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.304876 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-446gh" podUID="5133ffab-cfbb-4b1d-a161-d88f42ef37ce" containerName="registry-server" containerID="cri-o://3a082d038a2524fac8bc987dd4406cce128a3bc2fcaad85adee60c938e979894" gracePeriod=2 Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.324801 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-ccgqx"] Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.394661 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0d261b9-8134-46ec-b510-63fce197ccff-combined-ca-bundle\") pod \"ovn-controller-metrics-5szsr\" (UID: \"b0d261b9-8134-46ec-b510-63fce197ccff\") " pod="openstack/ovn-controller-metrics-5szsr" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.394733 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j46xl\" (UniqueName: \"kubernetes.io/projected/b0d261b9-8134-46ec-b510-63fce197ccff-kube-api-access-j46xl\") pod \"ovn-controller-metrics-5szsr\" (UID: \"b0d261b9-8134-46ec-b510-63fce197ccff\") " pod="openstack/ovn-controller-metrics-5szsr" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.394766 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/b0d261b9-8134-46ec-b510-63fce197ccff-ovn-rundir\") pod \"ovn-controller-metrics-5szsr\" (UID: \"b0d261b9-8134-46ec-b510-63fce197ccff\") " pod="openstack/ovn-controller-metrics-5szsr" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.394904 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/b0d261b9-8134-46ec-b510-63fce197ccff-ovs-rundir\") pod \"ovn-controller-metrics-5szsr\" (UID: \"b0d261b9-8134-46ec-b510-63fce197ccff\") " pod="openstack/ovn-controller-metrics-5szsr" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.394967 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0d261b9-8134-46ec-b510-63fce197ccff-config\") pod \"ovn-controller-metrics-5szsr\" (UID: \"b0d261b9-8134-46ec-b510-63fce197ccff\") " pod="openstack/ovn-controller-metrics-5szsr" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.394988 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0d261b9-8134-46ec-b510-63fce197ccff-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-5szsr\" (UID: \"b0d261b9-8134-46ec-b510-63fce197ccff\") " pod="openstack/ovn-controller-metrics-5szsr" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.399197 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/b0d261b9-8134-46ec-b510-63fce197ccff-ovn-rundir\") pod \"ovn-controller-metrics-5szsr\" (UID: \"b0d261b9-8134-46ec-b510-63fce197ccff\") " pod="openstack/ovn-controller-metrics-5szsr" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.399335 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/b0d261b9-8134-46ec-b510-63fce197ccff-ovs-rundir\") pod \"ovn-controller-metrics-5szsr\" (UID: \"b0d261b9-8134-46ec-b510-63fce197ccff\") " pod="openstack/ovn-controller-metrics-5szsr" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.400922 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0d261b9-8134-46ec-b510-63fce197ccff-config\") pod \"ovn-controller-metrics-5szsr\" (UID: \"b0d261b9-8134-46ec-b510-63fce197ccff\") " pod="openstack/ovn-controller-metrics-5szsr" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.403041 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0d261b9-8134-46ec-b510-63fce197ccff-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-5szsr\" (UID: \"b0d261b9-8134-46ec-b510-63fce197ccff\") " pod="openstack/ovn-controller-metrics-5szsr" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.405004 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0d261b9-8134-46ec-b510-63fce197ccff-combined-ca-bundle\") pod \"ovn-controller-metrics-5szsr\" (UID: \"b0d261b9-8134-46ec-b510-63fce197ccff\") " pod="openstack/ovn-controller-metrics-5szsr" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.455406 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j46xl\" (UniqueName: \"kubernetes.io/projected/b0d261b9-8134-46ec-b510-63fce197ccff-kube-api-access-j46xl\") pod \"ovn-controller-metrics-5szsr\" (UID: \"b0d261b9-8134-46ec-b510-63fce197ccff\") " pod="openstack/ovn-controller-metrics-5szsr" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.467599 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-5szsr" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.484889 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-554d4"] Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.486773 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-554d4" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.493107 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.557918 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-554d4"] Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.603257 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l96rk\" (UniqueName: \"kubernetes.io/projected/6b3734a5-609e-4ffc-ac55-1d9fb1e699cc-kube-api-access-l96rk\") pod \"dnsmasq-dns-7fd796d7df-554d4\" (UID: \"6b3734a5-609e-4ffc-ac55-1d9fb1e699cc\") " pod="openstack/dnsmasq-dns-7fd796d7df-554d4" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.603364 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b3734a5-609e-4ffc-ac55-1d9fb1e699cc-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-554d4\" (UID: \"6b3734a5-609e-4ffc-ac55-1d9fb1e699cc\") " pod="openstack/dnsmasq-dns-7fd796d7df-554d4" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.603407 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b3734a5-609e-4ffc-ac55-1d9fb1e699cc-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-554d4\" (UID: \"6b3734a5-609e-4ffc-ac55-1d9fb1e699cc\") " pod="openstack/dnsmasq-dns-7fd796d7df-554d4" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.603481 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b3734a5-609e-4ffc-ac55-1d9fb1e699cc-config\") pod \"dnsmasq-dns-7fd796d7df-554d4\" (UID: \"6b3734a5-609e-4ffc-ac55-1d9fb1e699cc\") " pod="openstack/dnsmasq-dns-7fd796d7df-554d4" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.615457 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-9q9hd"] Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.664194 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-dbv6d"] Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.667023 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-dbv6d" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.680207 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.692754 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-dbv6d"] Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.704851 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b3734a5-609e-4ffc-ac55-1d9fb1e699cc-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-554d4\" (UID: \"6b3734a5-609e-4ffc-ac55-1d9fb1e699cc\") " pod="openstack/dnsmasq-dns-7fd796d7df-554d4" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.704908 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b3734a5-609e-4ffc-ac55-1d9fb1e699cc-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-554d4\" (UID: \"6b3734a5-609e-4ffc-ac55-1d9fb1e699cc\") " pod="openstack/dnsmasq-dns-7fd796d7df-554d4" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.704964 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b3734a5-609e-4ffc-ac55-1d9fb1e699cc-config\") pod \"dnsmasq-dns-7fd796d7df-554d4\" (UID: \"6b3734a5-609e-4ffc-ac55-1d9fb1e699cc\") " pod="openstack/dnsmasq-dns-7fd796d7df-554d4" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.705030 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l96rk\" (UniqueName: \"kubernetes.io/projected/6b3734a5-609e-4ffc-ac55-1d9fb1e699cc-kube-api-access-l96rk\") pod \"dnsmasq-dns-7fd796d7df-554d4\" (UID: \"6b3734a5-609e-4ffc-ac55-1d9fb1e699cc\") " pod="openstack/dnsmasq-dns-7fd796d7df-554d4" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.706404 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b3734a5-609e-4ffc-ac55-1d9fb1e699cc-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-554d4\" (UID: \"6b3734a5-609e-4ffc-ac55-1d9fb1e699cc\") " pod="openstack/dnsmasq-dns-7fd796d7df-554d4" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.706997 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b3734a5-609e-4ffc-ac55-1d9fb1e699cc-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-554d4\" (UID: \"6b3734a5-609e-4ffc-ac55-1d9fb1e699cc\") " pod="openstack/dnsmasq-dns-7fd796d7df-554d4" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.707738 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b3734a5-609e-4ffc-ac55-1d9fb1e699cc-config\") pod \"dnsmasq-dns-7fd796d7df-554d4\" (UID: \"6b3734a5-609e-4ffc-ac55-1d9fb1e699cc\") " pod="openstack/dnsmasq-dns-7fd796d7df-554d4" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.746041 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l96rk\" (UniqueName: \"kubernetes.io/projected/6b3734a5-609e-4ffc-ac55-1d9fb1e699cc-kube-api-access-l96rk\") pod \"dnsmasq-dns-7fd796d7df-554d4\" (UID: \"6b3734a5-609e-4ffc-ac55-1d9fb1e699cc\") " pod="openstack/dnsmasq-dns-7fd796d7df-554d4" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.807036 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njqgn\" (UniqueName: \"kubernetes.io/projected/82f8e690-e10d-4852-bc45-beba0e5b666e-kube-api-access-njqgn\") pod \"dnsmasq-dns-86db49b7ff-dbv6d\" (UID: \"82f8e690-e10d-4852-bc45-beba0e5b666e\") " pod="openstack/dnsmasq-dns-86db49b7ff-dbv6d" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.807128 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/82f8e690-e10d-4852-bc45-beba0e5b666e-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-dbv6d\" (UID: \"82f8e690-e10d-4852-bc45-beba0e5b666e\") " pod="openstack/dnsmasq-dns-86db49b7ff-dbv6d" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.807315 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/82f8e690-e10d-4852-bc45-beba0e5b666e-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-dbv6d\" (UID: \"82f8e690-e10d-4852-bc45-beba0e5b666e\") " pod="openstack/dnsmasq-dns-86db49b7ff-dbv6d" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.807356 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82f8e690-e10d-4852-bc45-beba0e5b666e-config\") pod \"dnsmasq-dns-86db49b7ff-dbv6d\" (UID: \"82f8e690-e10d-4852-bc45-beba0e5b666e\") " pod="openstack/dnsmasq-dns-86db49b7ff-dbv6d" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.807410 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/82f8e690-e10d-4852-bc45-beba0e5b666e-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-dbv6d\" (UID: \"82f8e690-e10d-4852-bc45-beba0e5b666e\") " pod="openstack/dnsmasq-dns-86db49b7ff-dbv6d" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.854230 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-554d4" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.909514 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/82f8e690-e10d-4852-bc45-beba0e5b666e-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-dbv6d\" (UID: \"82f8e690-e10d-4852-bc45-beba0e5b666e\") " pod="openstack/dnsmasq-dns-86db49b7ff-dbv6d" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.909587 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82f8e690-e10d-4852-bc45-beba0e5b666e-config\") pod \"dnsmasq-dns-86db49b7ff-dbv6d\" (UID: \"82f8e690-e10d-4852-bc45-beba0e5b666e\") " pod="openstack/dnsmasq-dns-86db49b7ff-dbv6d" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.909634 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/82f8e690-e10d-4852-bc45-beba0e5b666e-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-dbv6d\" (UID: \"82f8e690-e10d-4852-bc45-beba0e5b666e\") " pod="openstack/dnsmasq-dns-86db49b7ff-dbv6d" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.909690 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njqgn\" (UniqueName: \"kubernetes.io/projected/82f8e690-e10d-4852-bc45-beba0e5b666e-kube-api-access-njqgn\") pod \"dnsmasq-dns-86db49b7ff-dbv6d\" (UID: \"82f8e690-e10d-4852-bc45-beba0e5b666e\") " pod="openstack/dnsmasq-dns-86db49b7ff-dbv6d" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.909737 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/82f8e690-e10d-4852-bc45-beba0e5b666e-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-dbv6d\" (UID: \"82f8e690-e10d-4852-bc45-beba0e5b666e\") " pod="openstack/dnsmasq-dns-86db49b7ff-dbv6d" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.910674 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/82f8e690-e10d-4852-bc45-beba0e5b666e-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-dbv6d\" (UID: \"82f8e690-e10d-4852-bc45-beba0e5b666e\") " pod="openstack/dnsmasq-dns-86db49b7ff-dbv6d" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.910755 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/82f8e690-e10d-4852-bc45-beba0e5b666e-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-dbv6d\" (UID: \"82f8e690-e10d-4852-bc45-beba0e5b666e\") " pod="openstack/dnsmasq-dns-86db49b7ff-dbv6d" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.910974 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/82f8e690-e10d-4852-bc45-beba0e5b666e-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-dbv6d\" (UID: \"82f8e690-e10d-4852-bc45-beba0e5b666e\") " pod="openstack/dnsmasq-dns-86db49b7ff-dbv6d" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.911034 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82f8e690-e10d-4852-bc45-beba0e5b666e-config\") pod \"dnsmasq-dns-86db49b7ff-dbv6d\" (UID: \"82f8e690-e10d-4852-bc45-beba0e5b666e\") " pod="openstack/dnsmasq-dns-86db49b7ff-dbv6d" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.927660 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njqgn\" (UniqueName: \"kubernetes.io/projected/82f8e690-e10d-4852-bc45-beba0e5b666e-kube-api-access-njqgn\") pod \"dnsmasq-dns-86db49b7ff-dbv6d\" (UID: \"82f8e690-e10d-4852-bc45-beba0e5b666e\") " pod="openstack/dnsmasq-dns-86db49b7ff-dbv6d" Nov 29 04:54:05 crc kubenswrapper[4799]: I1129 04:54:05.984963 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-dbv6d" Nov 29 04:54:07 crc kubenswrapper[4799]: I1129 04:54:07.196977 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 29 04:54:07 crc kubenswrapper[4799]: I1129 04:54:07.497776 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 04:54:07 crc kubenswrapper[4799]: I1129 04:54:07.497981 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 04:54:08 crc kubenswrapper[4799]: I1129 04:54:08.318041 4799 generic.go:334] "Generic (PLEG): container finished" podID="5133ffab-cfbb-4b1d-a161-d88f42ef37ce" containerID="3a082d038a2524fac8bc987dd4406cce128a3bc2fcaad85adee60c938e979894" exitCode=0 Nov 29 04:54:08 crc kubenswrapper[4799]: I1129 04:54:08.318867 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-446gh" event={"ID":"5133ffab-cfbb-4b1d-a161-d88f42ef37ce","Type":"ContainerDied","Data":"3a082d038a2524fac8bc987dd4406cce128a3bc2fcaad85adee60c938e979894"} Nov 29 04:54:09 crc kubenswrapper[4799]: I1129 04:54:09.393784 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7gtsn" Nov 29 04:54:10 crc kubenswrapper[4799]: I1129 04:54:10.492094 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7gtsn"] Nov 29 04:54:10 crc kubenswrapper[4799]: I1129 04:54:10.492661 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7gtsn" podUID="80a2712a-3766-49b1-bab4-a806bb44e21f" containerName="registry-server" containerID="cri-o://44dd2618c30517ff8ffaf8fc851a2767b7bf67324399a981f948ffecae837bcc" gracePeriod=2 Nov 29 04:54:11 crc kubenswrapper[4799]: I1129 04:54:11.344835 4799 generic.go:334] "Generic (PLEG): container finished" podID="80a2712a-3766-49b1-bab4-a806bb44e21f" containerID="44dd2618c30517ff8ffaf8fc851a2767b7bf67324399a981f948ffecae837bcc" exitCode=0 Nov 29 04:54:11 crc kubenswrapper[4799]: I1129 04:54:11.344888 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7gtsn" event={"ID":"80a2712a-3766-49b1-bab4-a806bb44e21f","Type":"ContainerDied","Data":"44dd2618c30517ff8ffaf8fc851a2767b7bf67324399a981f948ffecae837bcc"} Nov 29 04:54:11 crc kubenswrapper[4799]: I1129 04:54:11.521953 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Nov 29 04:54:11 crc kubenswrapper[4799]: I1129 04:54:11.626329 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Nov 29 04:54:12 crc kubenswrapper[4799]: I1129 04:54:12.902690 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-24kgw"] Nov 29 04:54:12 crc kubenswrapper[4799]: I1129 04:54:12.905053 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-24kgw" Nov 29 04:54:12 crc kubenswrapper[4799]: I1129 04:54:12.918871 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-24kgw"] Nov 29 04:54:13 crc kubenswrapper[4799]: I1129 04:54:13.047745 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36215eda-2af7-4388-992f-b31807c9d423-catalog-content\") pod \"redhat-marketplace-24kgw\" (UID: \"36215eda-2af7-4388-992f-b31807c9d423\") " pod="openshift-marketplace/redhat-marketplace-24kgw" Nov 29 04:54:13 crc kubenswrapper[4799]: I1129 04:54:13.047959 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36215eda-2af7-4388-992f-b31807c9d423-utilities\") pod \"redhat-marketplace-24kgw\" (UID: \"36215eda-2af7-4388-992f-b31807c9d423\") " pod="openshift-marketplace/redhat-marketplace-24kgw" Nov 29 04:54:13 crc kubenswrapper[4799]: I1129 04:54:13.048003 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7cv7\" (UniqueName: \"kubernetes.io/projected/36215eda-2af7-4388-992f-b31807c9d423-kube-api-access-h7cv7\") pod \"redhat-marketplace-24kgw\" (UID: \"36215eda-2af7-4388-992f-b31807c9d423\") " pod="openshift-marketplace/redhat-marketplace-24kgw" Nov 29 04:54:13 crc kubenswrapper[4799]: I1129 04:54:13.149925 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36215eda-2af7-4388-992f-b31807c9d423-utilities\") pod \"redhat-marketplace-24kgw\" (UID: \"36215eda-2af7-4388-992f-b31807c9d423\") " pod="openshift-marketplace/redhat-marketplace-24kgw" Nov 29 04:54:13 crc kubenswrapper[4799]: I1129 04:54:13.150007 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7cv7\" (UniqueName: \"kubernetes.io/projected/36215eda-2af7-4388-992f-b31807c9d423-kube-api-access-h7cv7\") pod \"redhat-marketplace-24kgw\" (UID: \"36215eda-2af7-4388-992f-b31807c9d423\") " pod="openshift-marketplace/redhat-marketplace-24kgw" Nov 29 04:54:13 crc kubenswrapper[4799]: I1129 04:54:13.150095 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36215eda-2af7-4388-992f-b31807c9d423-catalog-content\") pod \"redhat-marketplace-24kgw\" (UID: \"36215eda-2af7-4388-992f-b31807c9d423\") " pod="openshift-marketplace/redhat-marketplace-24kgw" Nov 29 04:54:13 crc kubenswrapper[4799]: I1129 04:54:13.151003 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36215eda-2af7-4388-992f-b31807c9d423-utilities\") pod \"redhat-marketplace-24kgw\" (UID: \"36215eda-2af7-4388-992f-b31807c9d423\") " pod="openshift-marketplace/redhat-marketplace-24kgw" Nov 29 04:54:13 crc kubenswrapper[4799]: I1129 04:54:13.151063 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36215eda-2af7-4388-992f-b31807c9d423-catalog-content\") pod \"redhat-marketplace-24kgw\" (UID: \"36215eda-2af7-4388-992f-b31807c9d423\") " pod="openshift-marketplace/redhat-marketplace-24kgw" Nov 29 04:54:13 crc kubenswrapper[4799]: I1129 04:54:13.170479 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7cv7\" (UniqueName: \"kubernetes.io/projected/36215eda-2af7-4388-992f-b31807c9d423-kube-api-access-h7cv7\") pod \"redhat-marketplace-24kgw\" (UID: \"36215eda-2af7-4388-992f-b31807c9d423\") " pod="openshift-marketplace/redhat-marketplace-24kgw" Nov 29 04:54:13 crc kubenswrapper[4799]: I1129 04:54:13.227179 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-24kgw" Nov 29 04:54:13 crc kubenswrapper[4799]: E1129 04:54:13.338749 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3a082d038a2524fac8bc987dd4406cce128a3bc2fcaad85adee60c938e979894 is running failed: container process not found" containerID="3a082d038a2524fac8bc987dd4406cce128a3bc2fcaad85adee60c938e979894" cmd=["grpc_health_probe","-addr=:50051"] Nov 29 04:54:13 crc kubenswrapper[4799]: E1129 04:54:13.339504 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3a082d038a2524fac8bc987dd4406cce128a3bc2fcaad85adee60c938e979894 is running failed: container process not found" containerID="3a082d038a2524fac8bc987dd4406cce128a3bc2fcaad85adee60c938e979894" cmd=["grpc_health_probe","-addr=:50051"] Nov 29 04:54:13 crc kubenswrapper[4799]: E1129 04:54:13.339883 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3a082d038a2524fac8bc987dd4406cce128a3bc2fcaad85adee60c938e979894 is running failed: container process not found" containerID="3a082d038a2524fac8bc987dd4406cce128a3bc2fcaad85adee60c938e979894" cmd=["grpc_health_probe","-addr=:50051"] Nov 29 04:54:13 crc kubenswrapper[4799]: E1129 04:54:13.340002 4799 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3a082d038a2524fac8bc987dd4406cce128a3bc2fcaad85adee60c938e979894 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-446gh" podUID="5133ffab-cfbb-4b1d-a161-d88f42ef37ce" containerName="registry-server" Nov 29 04:54:15 crc kubenswrapper[4799]: E1129 04:54:15.503219 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified" Nov 29 04:54:15 crc kubenswrapper[4799]: E1129 04:54:15.503828 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:openstack-network-exporter,Image:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,Command:[/app/openstack-network-exporter],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:OPENSTACK_NETWORK_EXPORTER_YAML,Value:/etc/config/openstack-network-exporter.yaml,ValueFrom:nil,},EnvVar{Name:CONFIG_HASH,Value:nf9hd7hc7h556h596hf7hb7h5b8hbfh555hbh57bh65bh557h96h84h5d7h68ch678h669h57chf7h576h5cch5f8hc9h646h6dh7ch64dh5d5h576q,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:ovsdb-rundir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovnmetrics.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovnmetrics.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-w4v6w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovsdbserver-nb-0_openstack(c79e4fc0-8786-40a6-a775-29a4767d072e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 29 04:54:15 crc kubenswrapper[4799]: E1129 04:54:15.506638 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovsdbserver-nb-0" podUID="c79e4fc0-8786-40a6-a775-29a4767d072e" Nov 29 04:54:15 crc kubenswrapper[4799]: E1129 04:54:15.512821 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified" Nov 29 04:54:15 crc kubenswrapper[4799]: E1129 04:54:15.513035 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:openstack-network-exporter,Image:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,Command:[/app/openstack-network-exporter],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:OPENSTACK_NETWORK_EXPORTER_YAML,Value:/etc/config/openstack-network-exporter.yaml,ValueFrom:nil,},EnvVar{Name:CONFIG_HASH,Value:n658h5b4h656h7bh5dbhc9h7ch5cch68dh555h5f7h75h5dbh5d5h675h575h66bhc4h5f8h576hb5h75h6ch548h5d6h5b8hf6h656hf5h695hfdh65q,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:ovsdb-rundir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovnmetrics.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovnmetrics.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-b276n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovsdbserver-sb-0_openstack(5bb0e45f-f281-49ec-8a11-37e1ace553db): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 29 04:54:15 crc kubenswrapper[4799]: E1129 04:54:15.514181 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovsdbserver-sb-0" podUID="5bb0e45f-f281-49ec-8a11-37e1ace553db" Nov 29 04:54:15 crc kubenswrapper[4799]: I1129 04:54:15.604785 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7gtsn" Nov 29 04:54:15 crc kubenswrapper[4799]: I1129 04:54:15.608688 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-446gh" Nov 29 04:54:15 crc kubenswrapper[4799]: I1129 04:54:15.696516 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5133ffab-cfbb-4b1d-a161-d88f42ef37ce-utilities\") pod \"5133ffab-cfbb-4b1d-a161-d88f42ef37ce\" (UID: \"5133ffab-cfbb-4b1d-a161-d88f42ef37ce\") " Nov 29 04:54:15 crc kubenswrapper[4799]: I1129 04:54:15.696997 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zdxb8\" (UniqueName: \"kubernetes.io/projected/5133ffab-cfbb-4b1d-a161-d88f42ef37ce-kube-api-access-zdxb8\") pod \"5133ffab-cfbb-4b1d-a161-d88f42ef37ce\" (UID: \"5133ffab-cfbb-4b1d-a161-d88f42ef37ce\") " Nov 29 04:54:15 crc kubenswrapper[4799]: I1129 04:54:15.697030 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n5gfl\" (UniqueName: \"kubernetes.io/projected/80a2712a-3766-49b1-bab4-a806bb44e21f-kube-api-access-n5gfl\") pod \"80a2712a-3766-49b1-bab4-a806bb44e21f\" (UID: \"80a2712a-3766-49b1-bab4-a806bb44e21f\") " Nov 29 04:54:15 crc kubenswrapper[4799]: I1129 04:54:15.697073 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80a2712a-3766-49b1-bab4-a806bb44e21f-catalog-content\") pod \"80a2712a-3766-49b1-bab4-a806bb44e21f\" (UID: \"80a2712a-3766-49b1-bab4-a806bb44e21f\") " Nov 29 04:54:15 crc kubenswrapper[4799]: I1129 04:54:15.697276 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80a2712a-3766-49b1-bab4-a806bb44e21f-utilities\") pod \"80a2712a-3766-49b1-bab4-a806bb44e21f\" (UID: \"80a2712a-3766-49b1-bab4-a806bb44e21f\") " Nov 29 04:54:15 crc kubenswrapper[4799]: I1129 04:54:15.697350 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5133ffab-cfbb-4b1d-a161-d88f42ef37ce-catalog-content\") pod \"5133ffab-cfbb-4b1d-a161-d88f42ef37ce\" (UID: \"5133ffab-cfbb-4b1d-a161-d88f42ef37ce\") " Nov 29 04:54:15 crc kubenswrapper[4799]: I1129 04:54:15.697684 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5133ffab-cfbb-4b1d-a161-d88f42ef37ce-utilities" (OuterVolumeSpecName: "utilities") pod "5133ffab-cfbb-4b1d-a161-d88f42ef37ce" (UID: "5133ffab-cfbb-4b1d-a161-d88f42ef37ce"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:54:15 crc kubenswrapper[4799]: I1129 04:54:15.697865 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5133ffab-cfbb-4b1d-a161-d88f42ef37ce-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:15 crc kubenswrapper[4799]: I1129 04:54:15.698484 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80a2712a-3766-49b1-bab4-a806bb44e21f-utilities" (OuterVolumeSpecName: "utilities") pod "80a2712a-3766-49b1-bab4-a806bb44e21f" (UID: "80a2712a-3766-49b1-bab4-a806bb44e21f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:54:15 crc kubenswrapper[4799]: I1129 04:54:15.701419 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80a2712a-3766-49b1-bab4-a806bb44e21f-kube-api-access-n5gfl" (OuterVolumeSpecName: "kube-api-access-n5gfl") pod "80a2712a-3766-49b1-bab4-a806bb44e21f" (UID: "80a2712a-3766-49b1-bab4-a806bb44e21f"). InnerVolumeSpecName "kube-api-access-n5gfl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:54:15 crc kubenswrapper[4799]: I1129 04:54:15.722288 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5133ffab-cfbb-4b1d-a161-d88f42ef37ce-kube-api-access-zdxb8" (OuterVolumeSpecName: "kube-api-access-zdxb8") pod "5133ffab-cfbb-4b1d-a161-d88f42ef37ce" (UID: "5133ffab-cfbb-4b1d-a161-d88f42ef37ce"). InnerVolumeSpecName "kube-api-access-zdxb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:54:15 crc kubenswrapper[4799]: I1129 04:54:15.771078 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80a2712a-3766-49b1-bab4-a806bb44e21f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "80a2712a-3766-49b1-bab4-a806bb44e21f" (UID: "80a2712a-3766-49b1-bab4-a806bb44e21f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:54:15 crc kubenswrapper[4799]: I1129 04:54:15.792337 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5133ffab-cfbb-4b1d-a161-d88f42ef37ce-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5133ffab-cfbb-4b1d-a161-d88f42ef37ce" (UID: "5133ffab-cfbb-4b1d-a161-d88f42ef37ce"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:54:15 crc kubenswrapper[4799]: I1129 04:54:15.799384 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80a2712a-3766-49b1-bab4-a806bb44e21f-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:15 crc kubenswrapper[4799]: I1129 04:54:15.799420 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5133ffab-cfbb-4b1d-a161-d88f42ef37ce-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:15 crc kubenswrapper[4799]: I1129 04:54:15.799434 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zdxb8\" (UniqueName: \"kubernetes.io/projected/5133ffab-cfbb-4b1d-a161-d88f42ef37ce-kube-api-access-zdxb8\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:15 crc kubenswrapper[4799]: I1129 04:54:15.799445 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n5gfl\" (UniqueName: \"kubernetes.io/projected/80a2712a-3766-49b1-bab4-a806bb44e21f-kube-api-access-n5gfl\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:15 crc kubenswrapper[4799]: I1129 04:54:15.799454 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80a2712a-3766-49b1-bab4-a806bb44e21f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:15 crc kubenswrapper[4799]: I1129 04:54:15.969411 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Nov 29 04:54:16 crc kubenswrapper[4799]: I1129 04:54:16.043634 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-5szsr"] Nov 29 04:54:16 crc kubenswrapper[4799]: I1129 04:54:16.097447 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-dbv6d"] Nov 29 04:54:16 crc kubenswrapper[4799]: W1129 04:54:16.104418 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod82f8e690_e10d_4852_bc45_beba0e5b666e.slice/crio-b8fa8d7f083f296493e244bc00d86d87ebbd00fe65007fdd24d4a61c202b0354 WatchSource:0}: Error finding container b8fa8d7f083f296493e244bc00d86d87ebbd00fe65007fdd24d4a61c202b0354: Status 404 returned error can't find the container with id b8fa8d7f083f296493e244bc00d86d87ebbd00fe65007fdd24d4a61c202b0354 Nov 29 04:54:16 crc kubenswrapper[4799]: I1129 04:54:16.120507 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-24kgw"] Nov 29 04:54:16 crc kubenswrapper[4799]: I1129 04:54:16.122937 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Nov 29 04:54:16 crc kubenswrapper[4799]: W1129 04:54:16.129435 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod36215eda_2af7_4388_992f_b31807c9d423.slice/crio-c20704c6b3ad06b103ea0f1de65eee4ad72a2ee7519799abbef097d50fe22743 WatchSource:0}: Error finding container c20704c6b3ad06b103ea0f1de65eee4ad72a2ee7519799abbef097d50fe22743: Status 404 returned error can't find the container with id c20704c6b3ad06b103ea0f1de65eee4ad72a2ee7519799abbef097d50fe22743 Nov 29 04:54:16 crc kubenswrapper[4799]: I1129 04:54:16.136995 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-554d4"] Nov 29 04:54:16 crc kubenswrapper[4799]: I1129 04:54:16.397425 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-dbv6d" event={"ID":"82f8e690-e10d-4852-bc45-beba0e5b666e","Type":"ContainerStarted","Data":"b8fa8d7f083f296493e244bc00d86d87ebbd00fe65007fdd24d4a61c202b0354"} Nov 29 04:54:16 crc kubenswrapper[4799]: I1129 04:54:16.400900 4799 generic.go:334] "Generic (PLEG): container finished" podID="f3627f09-e94d-4a00-94a2-7dcf7dfe7800" containerID="2a93c38935dd242dc36d68ba2d51dafe3cfbfb0cbfec76f2595dfdf0630e9f35" exitCode=0 Nov 29 04:54:16 crc kubenswrapper[4799]: I1129 04:54:16.400959 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-ccgqx" event={"ID":"f3627f09-e94d-4a00-94a2-7dcf7dfe7800","Type":"ContainerDied","Data":"2a93c38935dd242dc36d68ba2d51dafe3cfbfb0cbfec76f2595dfdf0630e9f35"} Nov 29 04:54:16 crc kubenswrapper[4799]: I1129 04:54:16.406393 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-24kgw" event={"ID":"36215eda-2af7-4388-992f-b31807c9d423","Type":"ContainerStarted","Data":"baba2e8023cbab4935c7039e780d17ac58c69198100aab1af217d7a18c88876f"} Nov 29 04:54:16 crc kubenswrapper[4799]: I1129 04:54:16.406517 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-24kgw" event={"ID":"36215eda-2af7-4388-992f-b31807c9d423","Type":"ContainerStarted","Data":"c20704c6b3ad06b103ea0f1de65eee4ad72a2ee7519799abbef097d50fe22743"} Nov 29 04:54:16 crc kubenswrapper[4799]: I1129 04:54:16.410940 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-5szsr" event={"ID":"b0d261b9-8134-46ec-b510-63fce197ccff","Type":"ContainerStarted","Data":"96d174d27f2d2b72ef836e6dfb52e28b20635b29981539f844d5f0df641eb2a3"} Nov 29 04:54:16 crc kubenswrapper[4799]: I1129 04:54:16.412325 4799 generic.go:334] "Generic (PLEG): container finished" podID="da2af6d0-1824-4820-84eb-0c491cb777a9" containerID="caa3c79c82037eadfc931ee2f6329e4e4e414aaf3688a3d6c655997ad24b6697" exitCode=0 Nov 29 04:54:16 crc kubenswrapper[4799]: I1129 04:54:16.412378 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-9q9hd" event={"ID":"da2af6d0-1824-4820-84eb-0c491cb777a9","Type":"ContainerDied","Data":"caa3c79c82037eadfc931ee2f6329e4e4e414aaf3688a3d6c655997ad24b6697"} Nov 29 04:54:16 crc kubenswrapper[4799]: I1129 04:54:16.421689 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-554d4" event={"ID":"6b3734a5-609e-4ffc-ac55-1d9fb1e699cc","Type":"ContainerStarted","Data":"aac3f393379d0d3c6033bba5698a20279fbce5d538785dac9c57c7e12b1fcb6c"} Nov 29 04:54:16 crc kubenswrapper[4799]: I1129 04:54:16.451660 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7gtsn" Nov 29 04:54:16 crc kubenswrapper[4799]: I1129 04:54:16.458830 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7gtsn" event={"ID":"80a2712a-3766-49b1-bab4-a806bb44e21f","Type":"ContainerDied","Data":"2eb02c5a5b0ad0d1dc3d1e4d223c8ebb4b48b957d13ef9aa3d784da5967431b0"} Nov 29 04:54:16 crc kubenswrapper[4799]: I1129 04:54:16.458899 4799 scope.go:117] "RemoveContainer" containerID="44dd2618c30517ff8ffaf8fc851a2767b7bf67324399a981f948ffecae837bcc" Nov 29 04:54:16 crc kubenswrapper[4799]: I1129 04:54:16.478642 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jpq6g" event={"ID":"89b0a912-606b-4d5d-bd40-09f89f2db649","Type":"ContainerStarted","Data":"7c548c9a584c5577fe8115c5f60db7f80c61da43605a9ba6dc27fd1fe751b204"} Nov 29 04:54:16 crc kubenswrapper[4799]: I1129 04:54:16.491368 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-446gh" Nov 29 04:54:16 crc kubenswrapper[4799]: I1129 04:54:16.492054 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-446gh" event={"ID":"5133ffab-cfbb-4b1d-a161-d88f42ef37ce","Type":"ContainerDied","Data":"773f0f87567ce61a58c274605115e369dbcb8919ed32e615e3d43f97b6543330"} Nov 29 04:54:16 crc kubenswrapper[4799]: E1129 04:54:16.495145 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"" pod="openstack/ovsdbserver-sb-0" podUID="5bb0e45f-f281-49ec-8a11-37e1ace553db" Nov 29 04:54:16 crc kubenswrapper[4799]: E1129 04:54:16.495349 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="c79e4fc0-8786-40a6-a775-29a4767d072e" Nov 29 04:54:16 crc kubenswrapper[4799]: I1129 04:54:16.530552 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jpq6g" podStartSLOduration=4.084414592 podStartE2EDuration="22.53042475s" podCreationTimestamp="2025-11-29 04:53:54 +0000 UTC" firstStartedPulling="2025-11-29 04:53:57.202185454 +0000 UTC m=+912.845115854" lastFinishedPulling="2025-11-29 04:54:15.648195612 +0000 UTC m=+931.291126012" observedRunningTime="2025-11-29 04:54:16.502214634 +0000 UTC m=+932.145145054" watchObservedRunningTime="2025-11-29 04:54:16.53042475 +0000 UTC m=+932.173355150" Nov 29 04:54:16 crc kubenswrapper[4799]: I1129 04:54:16.560600 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7gtsn"] Nov 29 04:54:16 crc kubenswrapper[4799]: I1129 04:54:16.567820 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7gtsn"] Nov 29 04:54:16 crc kubenswrapper[4799]: I1129 04:54:16.605451 4799 scope.go:117] "RemoveContainer" containerID="401ec315c842097520dc3a517a3d7f83b9ed67273f265d705da6318a2b3e540c" Nov 29 04:54:16 crc kubenswrapper[4799]: I1129 04:54:16.618646 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-446gh"] Nov 29 04:54:16 crc kubenswrapper[4799]: I1129 04:54:16.625997 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-446gh"] Nov 29 04:54:16 crc kubenswrapper[4799]: I1129 04:54:16.663077 4799 scope.go:117] "RemoveContainer" containerID="148179253a550e8c180d17b00014302c25693e63366864ce710dc422e93e0425" Nov 29 04:54:16 crc kubenswrapper[4799]: I1129 04:54:16.692321 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5133ffab-cfbb-4b1d-a161-d88f42ef37ce" path="/var/lib/kubelet/pods/5133ffab-cfbb-4b1d-a161-d88f42ef37ce/volumes" Nov 29 04:54:16 crc kubenswrapper[4799]: I1129 04:54:16.692977 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80a2712a-3766-49b1-bab4-a806bb44e21f" path="/var/lib/kubelet/pods/80a2712a-3766-49b1-bab4-a806bb44e21f/volumes" Nov 29 04:54:16 crc kubenswrapper[4799]: I1129 04:54:16.772699 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-ccgqx" Nov 29 04:54:16 crc kubenswrapper[4799]: I1129 04:54:16.783014 4799 scope.go:117] "RemoveContainer" containerID="3a082d038a2524fac8bc987dd4406cce128a3bc2fcaad85adee60c938e979894" Nov 29 04:54:16 crc kubenswrapper[4799]: I1129 04:54:16.808992 4799 scope.go:117] "RemoveContainer" containerID="333991016bafcc9c6d8c9cafe90b47b91f9226151db9ecd3b2956128102352ec" Nov 29 04:54:16 crc kubenswrapper[4799]: I1129 04:54:16.822587 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3627f09-e94d-4a00-94a2-7dcf7dfe7800-config\") pod \"f3627f09-e94d-4a00-94a2-7dcf7dfe7800\" (UID: \"f3627f09-e94d-4a00-94a2-7dcf7dfe7800\") " Nov 29 04:54:16 crc kubenswrapper[4799]: I1129 04:54:16.822747 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f3627f09-e94d-4a00-94a2-7dcf7dfe7800-dns-svc\") pod \"f3627f09-e94d-4a00-94a2-7dcf7dfe7800\" (UID: \"f3627f09-e94d-4a00-94a2-7dcf7dfe7800\") " Nov 29 04:54:16 crc kubenswrapper[4799]: I1129 04:54:16.822822 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crcdk\" (UniqueName: \"kubernetes.io/projected/f3627f09-e94d-4a00-94a2-7dcf7dfe7800-kube-api-access-crcdk\") pod \"f3627f09-e94d-4a00-94a2-7dcf7dfe7800\" (UID: \"f3627f09-e94d-4a00-94a2-7dcf7dfe7800\") " Nov 29 04:54:16 crc kubenswrapper[4799]: I1129 04:54:16.836925 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3627f09-e94d-4a00-94a2-7dcf7dfe7800-kube-api-access-crcdk" (OuterVolumeSpecName: "kube-api-access-crcdk") pod "f3627f09-e94d-4a00-94a2-7dcf7dfe7800" (UID: "f3627f09-e94d-4a00-94a2-7dcf7dfe7800"). InnerVolumeSpecName "kube-api-access-crcdk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:54:16 crc kubenswrapper[4799]: I1129 04:54:16.842004 4799 scope.go:117] "RemoveContainer" containerID="0132be20cfe9fdf4df45dedb92807009b0008025be3bc156101b9e65e89a2063" Nov 29 04:54:16 crc kubenswrapper[4799]: I1129 04:54:16.848419 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f3627f09-e94d-4a00-94a2-7dcf7dfe7800-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f3627f09-e94d-4a00-94a2-7dcf7dfe7800" (UID: "f3627f09-e94d-4a00-94a2-7dcf7dfe7800"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:54:16 crc kubenswrapper[4799]: I1129 04:54:16.861613 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Nov 29 04:54:16 crc kubenswrapper[4799]: I1129 04:54:16.864297 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f3627f09-e94d-4a00-94a2-7dcf7dfe7800-config" (OuterVolumeSpecName: "config") pod "f3627f09-e94d-4a00-94a2-7dcf7dfe7800" (UID: "f3627f09-e94d-4a00-94a2-7dcf7dfe7800"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:54:16 crc kubenswrapper[4799]: I1129 04:54:16.867909 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-9q9hd" Nov 29 04:54:16 crc kubenswrapper[4799]: I1129 04:54:16.924834 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da2af6d0-1824-4820-84eb-0c491cb777a9-dns-svc\") pod \"da2af6d0-1824-4820-84eb-0c491cb777a9\" (UID: \"da2af6d0-1824-4820-84eb-0c491cb777a9\") " Nov 29 04:54:16 crc kubenswrapper[4799]: I1129 04:54:16.924900 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ccqck\" (UniqueName: \"kubernetes.io/projected/da2af6d0-1824-4820-84eb-0c491cb777a9-kube-api-access-ccqck\") pod \"da2af6d0-1824-4820-84eb-0c491cb777a9\" (UID: \"da2af6d0-1824-4820-84eb-0c491cb777a9\") " Nov 29 04:54:16 crc kubenswrapper[4799]: I1129 04:54:16.924924 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da2af6d0-1824-4820-84eb-0c491cb777a9-config\") pod \"da2af6d0-1824-4820-84eb-0c491cb777a9\" (UID: \"da2af6d0-1824-4820-84eb-0c491cb777a9\") " Nov 29 04:54:16 crc kubenswrapper[4799]: I1129 04:54:16.925363 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f3627f09-e94d-4a00-94a2-7dcf7dfe7800-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:16 crc kubenswrapper[4799]: I1129 04:54:16.925388 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crcdk\" (UniqueName: \"kubernetes.io/projected/f3627f09-e94d-4a00-94a2-7dcf7dfe7800-kube-api-access-crcdk\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:16 crc kubenswrapper[4799]: I1129 04:54:16.925400 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3627f09-e94d-4a00-94a2-7dcf7dfe7800-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:16 crc kubenswrapper[4799]: I1129 04:54:16.933658 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da2af6d0-1824-4820-84eb-0c491cb777a9-kube-api-access-ccqck" (OuterVolumeSpecName: "kube-api-access-ccqck") pod "da2af6d0-1824-4820-84eb-0c491cb777a9" (UID: "da2af6d0-1824-4820-84eb-0c491cb777a9"). InnerVolumeSpecName "kube-api-access-ccqck". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:54:16 crc kubenswrapper[4799]: I1129 04:54:16.951282 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da2af6d0-1824-4820-84eb-0c491cb777a9-config" (OuterVolumeSpecName: "config") pod "da2af6d0-1824-4820-84eb-0c491cb777a9" (UID: "da2af6d0-1824-4820-84eb-0c491cb777a9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:54:16 crc kubenswrapper[4799]: I1129 04:54:16.953667 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da2af6d0-1824-4820-84eb-0c491cb777a9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "da2af6d0-1824-4820-84eb-0c491cb777a9" (UID: "da2af6d0-1824-4820-84eb-0c491cb777a9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:54:17 crc kubenswrapper[4799]: I1129 04:54:17.027654 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da2af6d0-1824-4820-84eb-0c491cb777a9-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:17 crc kubenswrapper[4799]: I1129 04:54:17.027744 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ccqck\" (UniqueName: \"kubernetes.io/projected/da2af6d0-1824-4820-84eb-0c491cb777a9-kube-api-access-ccqck\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:17 crc kubenswrapper[4799]: I1129 04:54:17.027760 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da2af6d0-1824-4820-84eb-0c491cb777a9-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:17 crc kubenswrapper[4799]: I1129 04:54:17.506082 4799 generic.go:334] "Generic (PLEG): container finished" podID="82f8e690-e10d-4852-bc45-beba0e5b666e" containerID="e0521ca520400ce0ca8b2930b538fc5edee5a488dc7c80d203302043be8fb969" exitCode=0 Nov 29 04:54:17 crc kubenswrapper[4799]: I1129 04:54:17.506192 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-dbv6d" event={"ID":"82f8e690-e10d-4852-bc45-beba0e5b666e","Type":"ContainerDied","Data":"e0521ca520400ce0ca8b2930b538fc5edee5a488dc7c80d203302043be8fb969"} Nov 29 04:54:17 crc kubenswrapper[4799]: I1129 04:54:17.510459 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-ccgqx" event={"ID":"f3627f09-e94d-4a00-94a2-7dcf7dfe7800","Type":"ContainerDied","Data":"02adf8c7c010cdf77d280123300275260a01c4c9ec438444c80d8a0c324b69a2"} Nov 29 04:54:17 crc kubenswrapper[4799]: I1129 04:54:17.510507 4799 scope.go:117] "RemoveContainer" containerID="2a93c38935dd242dc36d68ba2d51dafe3cfbfb0cbfec76f2595dfdf0630e9f35" Nov 29 04:54:17 crc kubenswrapper[4799]: I1129 04:54:17.510664 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-ccgqx" Nov 29 04:54:17 crc kubenswrapper[4799]: I1129 04:54:17.535071 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-9q9hd" Nov 29 04:54:17 crc kubenswrapper[4799]: I1129 04:54:17.534989 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-9q9hd" event={"ID":"da2af6d0-1824-4820-84eb-0c491cb777a9","Type":"ContainerDied","Data":"ad8249dfae367d7e0f5593bfe8683dd96de7e32947539f1423cba28b54dbc50a"} Nov 29 04:54:17 crc kubenswrapper[4799]: I1129 04:54:17.543159 4799 generic.go:334] "Generic (PLEG): container finished" podID="36215eda-2af7-4388-992f-b31807c9d423" containerID="baba2e8023cbab4935c7039e780d17ac58c69198100aab1af217d7a18c88876f" exitCode=0 Nov 29 04:54:17 crc kubenswrapper[4799]: I1129 04:54:17.543349 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-24kgw" event={"ID":"36215eda-2af7-4388-992f-b31807c9d423","Type":"ContainerDied","Data":"baba2e8023cbab4935c7039e780d17ac58c69198100aab1af217d7a18c88876f"} Nov 29 04:54:17 crc kubenswrapper[4799]: I1129 04:54:17.565328 4799 generic.go:334] "Generic (PLEG): container finished" podID="6b3734a5-609e-4ffc-ac55-1d9fb1e699cc" containerID="e91847d686e65cefcb856dfa8a8fd1a0b792ecec149e99d15246b3fc74fd2a63" exitCode=0 Nov 29 04:54:17 crc kubenswrapper[4799]: I1129 04:54:17.565543 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-554d4" event={"ID":"6b3734a5-609e-4ffc-ac55-1d9fb1e699cc","Type":"ContainerDied","Data":"e91847d686e65cefcb856dfa8a8fd1a0b792ecec149e99d15246b3fc74fd2a63"} Nov 29 04:54:17 crc kubenswrapper[4799]: I1129 04:54:17.572745 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-5szsr" event={"ID":"b0d261b9-8134-46ec-b510-63fce197ccff","Type":"ContainerStarted","Data":"f016a14868b645cb6bfbd673631f7d96942f98b98e5fdf37272ca0fa29d9844a"} Nov 29 04:54:17 crc kubenswrapper[4799]: I1129 04:54:17.651586 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-5szsr" podStartSLOduration=12.035823058 podStartE2EDuration="12.651547681s" podCreationTimestamp="2025-11-29 04:54:05 +0000 UTC" firstStartedPulling="2025-11-29 04:54:16.047702968 +0000 UTC m=+931.690633368" lastFinishedPulling="2025-11-29 04:54:16.663427591 +0000 UTC m=+932.306357991" observedRunningTime="2025-11-29 04:54:17.605477362 +0000 UTC m=+933.248407772" watchObservedRunningTime="2025-11-29 04:54:17.651547681 +0000 UTC m=+933.294478081" Nov 29 04:54:17 crc kubenswrapper[4799]: I1129 04:54:17.690371 4799 scope.go:117] "RemoveContainer" containerID="caa3c79c82037eadfc931ee2f6329e4e4e414aaf3688a3d6c655997ad24b6697" Nov 29 04:54:17 crc kubenswrapper[4799]: I1129 04:54:17.786239 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-9q9hd"] Nov 29 04:54:17 crc kubenswrapper[4799]: I1129 04:54:17.794555 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-9q9hd"] Nov 29 04:54:17 crc kubenswrapper[4799]: I1129 04:54:17.817864 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-ccgqx"] Nov 29 04:54:17 crc kubenswrapper[4799]: I1129 04:54:17.831320 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-ccgqx"] Nov 29 04:54:18 crc kubenswrapper[4799]: I1129 04:54:18.587920 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-dbv6d" event={"ID":"82f8e690-e10d-4852-bc45-beba0e5b666e","Type":"ContainerStarted","Data":"7d8eadeceea35e76b407cce3a32ef5e187792093f406f2a855c2100538b72f7b"} Nov 29 04:54:18 crc kubenswrapper[4799]: I1129 04:54:18.590035 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-dbv6d" Nov 29 04:54:18 crc kubenswrapper[4799]: I1129 04:54:18.595602 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"c79e4fc0-8786-40a6-a775-29a4767d072e","Type":"ContainerStarted","Data":"59b6bc1d6010e11d7b709462d7310411a183a71af2b7868fe03d92ee833cf13c"} Nov 29 04:54:18 crc kubenswrapper[4799]: I1129 04:54:18.599454 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-554d4" event={"ID":"6b3734a5-609e-4ffc-ac55-1d9fb1e699cc","Type":"ContainerStarted","Data":"662675aecdd4c6f6a80b63ef0c29e8aaa6b5b4687e8314e732545cbd84a8e52f"} Nov 29 04:54:18 crc kubenswrapper[4799]: I1129 04:54:18.599516 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7fd796d7df-554d4" Nov 29 04:54:18 crc kubenswrapper[4799]: I1129 04:54:18.623967 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-dbv6d" podStartSLOduration=13.623933897 podStartE2EDuration="13.623933897s" podCreationTimestamp="2025-11-29 04:54:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:54:18.611634688 +0000 UTC m=+934.254565088" watchObservedRunningTime="2025-11-29 04:54:18.623933897 +0000 UTC m=+934.266864297" Nov 29 04:54:18 crc kubenswrapper[4799]: I1129 04:54:18.638638 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=49.414068345 podStartE2EDuration="58.638612155s" podCreationTimestamp="2025-11-29 04:53:20 +0000 UTC" firstStartedPulling="2025-11-29 04:53:44.926826924 +0000 UTC m=+900.569757324" lastFinishedPulling="2025-11-29 04:53:54.151370734 +0000 UTC m=+909.794301134" observedRunningTime="2025-11-29 04:54:18.63292216 +0000 UTC m=+934.275852560" watchObservedRunningTime="2025-11-29 04:54:18.638612155 +0000 UTC m=+934.281542545" Nov 29 04:54:18 crc kubenswrapper[4799]: I1129 04:54:18.662227 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7fd796d7df-554d4" podStartSLOduration=13.662193382 podStartE2EDuration="13.662193382s" podCreationTimestamp="2025-11-29 04:54:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:54:18.65661421 +0000 UTC m=+934.299544640" watchObservedRunningTime="2025-11-29 04:54:18.662193382 +0000 UTC m=+934.305123782" Nov 29 04:54:18 crc kubenswrapper[4799]: I1129 04:54:18.673028 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da2af6d0-1824-4820-84eb-0c491cb777a9" path="/var/lib/kubelet/pods/da2af6d0-1824-4820-84eb-0c491cb777a9/volumes" Nov 29 04:54:18 crc kubenswrapper[4799]: I1129 04:54:18.673884 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3627f09-e94d-4a00-94a2-7dcf7dfe7800" path="/var/lib/kubelet/pods/f3627f09-e94d-4a00-94a2-7dcf7dfe7800/volumes" Nov 29 04:54:18 crc kubenswrapper[4799]: I1129 04:54:18.821499 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Nov 29 04:54:18 crc kubenswrapper[4799]: I1129 04:54:18.859506 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Nov 29 04:54:18 crc kubenswrapper[4799]: I1129 04:54:18.880372 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Nov 29 04:54:18 crc kubenswrapper[4799]: I1129 04:54:18.911084 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Nov 29 04:54:19 crc kubenswrapper[4799]: I1129 04:54:19.608185 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"5bb0e45f-f281-49ec-8a11-37e1ace553db","Type":"ContainerStarted","Data":"74d89ee44b4fbc1ab7f4ebf43510a0af44df5b5311e28c7ba50282c1879a8633"} Nov 29 04:54:19 crc kubenswrapper[4799]: I1129 04:54:19.608387 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Nov 29 04:54:19 crc kubenswrapper[4799]: I1129 04:54:19.612067 4799 generic.go:334] "Generic (PLEG): container finished" podID="36215eda-2af7-4388-992f-b31807c9d423" containerID="484a1d6e238a655ab1bf8dc288912082c32099706e4b71564d8a42fc9796e7b8" exitCode=0 Nov 29 04:54:19 crc kubenswrapper[4799]: I1129 04:54:19.612204 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-24kgw" event={"ID":"36215eda-2af7-4388-992f-b31807c9d423","Type":"ContainerDied","Data":"484a1d6e238a655ab1bf8dc288912082c32099706e4b71564d8a42fc9796e7b8"} Nov 29 04:54:19 crc kubenswrapper[4799]: I1129 04:54:19.636239 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=48.532191533 podStartE2EDuration="56.636210087s" podCreationTimestamp="2025-11-29 04:53:23 +0000 UTC" firstStartedPulling="2025-11-29 04:53:45.855940339 +0000 UTC m=+901.498870739" lastFinishedPulling="2025-11-29 04:53:53.959958893 +0000 UTC m=+909.602889293" observedRunningTime="2025-11-29 04:54:19.629601922 +0000 UTC m=+935.272532322" watchObservedRunningTime="2025-11-29 04:54:19.636210087 +0000 UTC m=+935.279140497" Nov 29 04:54:19 crc kubenswrapper[4799]: I1129 04:54:19.670407 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Nov 29 04:54:19 crc kubenswrapper[4799]: I1129 04:54:19.670498 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.029677 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Nov 29 04:54:20 crc kubenswrapper[4799]: E1129 04:54:20.030424 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5133ffab-cfbb-4b1d-a161-d88f42ef37ce" containerName="extract-utilities" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.030449 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5133ffab-cfbb-4b1d-a161-d88f42ef37ce" containerName="extract-utilities" Nov 29 04:54:20 crc kubenswrapper[4799]: E1129 04:54:20.030462 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5133ffab-cfbb-4b1d-a161-d88f42ef37ce" containerName="registry-server" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.030471 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5133ffab-cfbb-4b1d-a161-d88f42ef37ce" containerName="registry-server" Nov 29 04:54:20 crc kubenswrapper[4799]: E1129 04:54:20.030490 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5133ffab-cfbb-4b1d-a161-d88f42ef37ce" containerName="extract-content" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.030502 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5133ffab-cfbb-4b1d-a161-d88f42ef37ce" containerName="extract-content" Nov 29 04:54:20 crc kubenswrapper[4799]: E1129 04:54:20.030523 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3627f09-e94d-4a00-94a2-7dcf7dfe7800" containerName="init" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.030533 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3627f09-e94d-4a00-94a2-7dcf7dfe7800" containerName="init" Nov 29 04:54:20 crc kubenswrapper[4799]: E1129 04:54:20.030548 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80a2712a-3766-49b1-bab4-a806bb44e21f" containerName="extract-content" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.030556 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="80a2712a-3766-49b1-bab4-a806bb44e21f" containerName="extract-content" Nov 29 04:54:20 crc kubenswrapper[4799]: E1129 04:54:20.030572 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da2af6d0-1824-4820-84eb-0c491cb777a9" containerName="init" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.030581 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="da2af6d0-1824-4820-84eb-0c491cb777a9" containerName="init" Nov 29 04:54:20 crc kubenswrapper[4799]: E1129 04:54:20.030601 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80a2712a-3766-49b1-bab4-a806bb44e21f" containerName="extract-utilities" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.030610 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="80a2712a-3766-49b1-bab4-a806bb44e21f" containerName="extract-utilities" Nov 29 04:54:20 crc kubenswrapper[4799]: E1129 04:54:20.030626 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80a2712a-3766-49b1-bab4-a806bb44e21f" containerName="registry-server" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.030634 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="80a2712a-3766-49b1-bab4-a806bb44e21f" containerName="registry-server" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.030868 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="5133ffab-cfbb-4b1d-a161-d88f42ef37ce" containerName="registry-server" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.030891 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="da2af6d0-1824-4820-84eb-0c491cb777a9" containerName="init" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.030903 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3627f09-e94d-4a00-94a2-7dcf7dfe7800" containerName="init" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.030916 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="80a2712a-3766-49b1-bab4-a806bb44e21f" containerName="registry-server" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.032051 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.036172 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-7hxz4" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.036549 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.036674 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.037806 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.062015 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.095760 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8c0babb-3163-4a72-a6d5-4fc5e64bee31-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"c8c0babb-3163-4a72-a6d5-4fc5e64bee31\") " pod="openstack/ovn-northd-0" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.096620 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8c0babb-3163-4a72-a6d5-4fc5e64bee31-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"c8c0babb-3163-4a72-a6d5-4fc5e64bee31\") " pod="openstack/ovn-northd-0" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.096672 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4rrz\" (UniqueName: \"kubernetes.io/projected/c8c0babb-3163-4a72-a6d5-4fc5e64bee31-kube-api-access-h4rrz\") pod \"ovn-northd-0\" (UID: \"c8c0babb-3163-4a72-a6d5-4fc5e64bee31\") " pod="openstack/ovn-northd-0" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.096699 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8c0babb-3163-4a72-a6d5-4fc5e64bee31-config\") pod \"ovn-northd-0\" (UID: \"c8c0babb-3163-4a72-a6d5-4fc5e64bee31\") " pod="openstack/ovn-northd-0" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.096723 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c8c0babb-3163-4a72-a6d5-4fc5e64bee31-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"c8c0babb-3163-4a72-a6d5-4fc5e64bee31\") " pod="openstack/ovn-northd-0" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.096758 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8c0babb-3163-4a72-a6d5-4fc5e64bee31-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"c8c0babb-3163-4a72-a6d5-4fc5e64bee31\") " pod="openstack/ovn-northd-0" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.096805 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c8c0babb-3163-4a72-a6d5-4fc5e64bee31-scripts\") pod \"ovn-northd-0\" (UID: \"c8c0babb-3163-4a72-a6d5-4fc5e64bee31\") " pod="openstack/ovn-northd-0" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.198293 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8c0babb-3163-4a72-a6d5-4fc5e64bee31-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"c8c0babb-3163-4a72-a6d5-4fc5e64bee31\") " pod="openstack/ovn-northd-0" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.198359 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c8c0babb-3163-4a72-a6d5-4fc5e64bee31-scripts\") pod \"ovn-northd-0\" (UID: \"c8c0babb-3163-4a72-a6d5-4fc5e64bee31\") " pod="openstack/ovn-northd-0" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.198413 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8c0babb-3163-4a72-a6d5-4fc5e64bee31-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"c8c0babb-3163-4a72-a6d5-4fc5e64bee31\") " pod="openstack/ovn-northd-0" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.198455 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8c0babb-3163-4a72-a6d5-4fc5e64bee31-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"c8c0babb-3163-4a72-a6d5-4fc5e64bee31\") " pod="openstack/ovn-northd-0" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.198488 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4rrz\" (UniqueName: \"kubernetes.io/projected/c8c0babb-3163-4a72-a6d5-4fc5e64bee31-kube-api-access-h4rrz\") pod \"ovn-northd-0\" (UID: \"c8c0babb-3163-4a72-a6d5-4fc5e64bee31\") " pod="openstack/ovn-northd-0" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.198512 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8c0babb-3163-4a72-a6d5-4fc5e64bee31-config\") pod \"ovn-northd-0\" (UID: \"c8c0babb-3163-4a72-a6d5-4fc5e64bee31\") " pod="openstack/ovn-northd-0" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.198537 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c8c0babb-3163-4a72-a6d5-4fc5e64bee31-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"c8c0babb-3163-4a72-a6d5-4fc5e64bee31\") " pod="openstack/ovn-northd-0" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.199144 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c8c0babb-3163-4a72-a6d5-4fc5e64bee31-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"c8c0babb-3163-4a72-a6d5-4fc5e64bee31\") " pod="openstack/ovn-northd-0" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.199661 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c8c0babb-3163-4a72-a6d5-4fc5e64bee31-scripts\") pod \"ovn-northd-0\" (UID: \"c8c0babb-3163-4a72-a6d5-4fc5e64bee31\") " pod="openstack/ovn-northd-0" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.199920 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8c0babb-3163-4a72-a6d5-4fc5e64bee31-config\") pod \"ovn-northd-0\" (UID: \"c8c0babb-3163-4a72-a6d5-4fc5e64bee31\") " pod="openstack/ovn-northd-0" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.205438 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8c0babb-3163-4a72-a6d5-4fc5e64bee31-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"c8c0babb-3163-4a72-a6d5-4fc5e64bee31\") " pod="openstack/ovn-northd-0" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.206254 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8c0babb-3163-4a72-a6d5-4fc5e64bee31-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"c8c0babb-3163-4a72-a6d5-4fc5e64bee31\") " pod="openstack/ovn-northd-0" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.208389 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8c0babb-3163-4a72-a6d5-4fc5e64bee31-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"c8c0babb-3163-4a72-a6d5-4fc5e64bee31\") " pod="openstack/ovn-northd-0" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.220028 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4rrz\" (UniqueName: \"kubernetes.io/projected/c8c0babb-3163-4a72-a6d5-4fc5e64bee31-kube-api-access-h4rrz\") pod \"ovn-northd-0\" (UID: \"c8c0babb-3163-4a72-a6d5-4fc5e64bee31\") " pod="openstack/ovn-northd-0" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.358630 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.636840 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-24kgw" event={"ID":"36215eda-2af7-4388-992f-b31807c9d423","Type":"ContainerStarted","Data":"fb79e68053b18642c9521ab4a168eee0942b1899c610dfa8021255d33f5d6189"} Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.664491 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-24kgw" podStartSLOduration=6.104148931 podStartE2EDuration="8.664467465s" podCreationTimestamp="2025-11-29 04:54:12 +0000 UTC" firstStartedPulling="2025-11-29 04:54:17.555531892 +0000 UTC m=+933.198462292" lastFinishedPulling="2025-11-29 04:54:20.115850426 +0000 UTC m=+935.758780826" observedRunningTime="2025-11-29 04:54:20.663962502 +0000 UTC m=+936.306892902" watchObservedRunningTime="2025-11-29 04:54:20.664467465 +0000 UTC m=+936.307397865" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.686381 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-s2hlx"] Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.693745 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-s2hlx" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.698384 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-5610-account-create-update-gwfrn"] Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.701545 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-5610-account-create-update-gwfrn" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.704266 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.714883 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-s2hlx"] Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.726866 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-5610-account-create-update-gwfrn"] Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.847004 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77sjh\" (UniqueName: \"kubernetes.io/projected/2d6648ca-f3df-4345-a2ac-d147e7ba85e1-kube-api-access-77sjh\") pod \"glance-5610-account-create-update-gwfrn\" (UID: \"2d6648ca-f3df-4345-a2ac-d147e7ba85e1\") " pod="openstack/glance-5610-account-create-update-gwfrn" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.847153 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d6648ca-f3df-4345-a2ac-d147e7ba85e1-operator-scripts\") pod \"glance-5610-account-create-update-gwfrn\" (UID: \"2d6648ca-f3df-4345-a2ac-d147e7ba85e1\") " pod="openstack/glance-5610-account-create-update-gwfrn" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.847276 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f356f37d-81e1-4554-abfe-8d13412ae662-operator-scripts\") pod \"glance-db-create-s2hlx\" (UID: \"f356f37d-81e1-4554-abfe-8d13412ae662\") " pod="openstack/glance-db-create-s2hlx" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.847325 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92rbj\" (UniqueName: \"kubernetes.io/projected/f356f37d-81e1-4554-abfe-8d13412ae662-kube-api-access-92rbj\") pod \"glance-db-create-s2hlx\" (UID: \"f356f37d-81e1-4554-abfe-8d13412ae662\") " pod="openstack/glance-db-create-s2hlx" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.892386 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.949121 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f356f37d-81e1-4554-abfe-8d13412ae662-operator-scripts\") pod \"glance-db-create-s2hlx\" (UID: \"f356f37d-81e1-4554-abfe-8d13412ae662\") " pod="openstack/glance-db-create-s2hlx" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.949192 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92rbj\" (UniqueName: \"kubernetes.io/projected/f356f37d-81e1-4554-abfe-8d13412ae662-kube-api-access-92rbj\") pod \"glance-db-create-s2hlx\" (UID: \"f356f37d-81e1-4554-abfe-8d13412ae662\") " pod="openstack/glance-db-create-s2hlx" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.949824 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77sjh\" (UniqueName: \"kubernetes.io/projected/2d6648ca-f3df-4345-a2ac-d147e7ba85e1-kube-api-access-77sjh\") pod \"glance-5610-account-create-update-gwfrn\" (UID: \"2d6648ca-f3df-4345-a2ac-d147e7ba85e1\") " pod="openstack/glance-5610-account-create-update-gwfrn" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.950071 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d6648ca-f3df-4345-a2ac-d147e7ba85e1-operator-scripts\") pod \"glance-5610-account-create-update-gwfrn\" (UID: \"2d6648ca-f3df-4345-a2ac-d147e7ba85e1\") " pod="openstack/glance-5610-account-create-update-gwfrn" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.950113 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f356f37d-81e1-4554-abfe-8d13412ae662-operator-scripts\") pod \"glance-db-create-s2hlx\" (UID: \"f356f37d-81e1-4554-abfe-8d13412ae662\") " pod="openstack/glance-db-create-s2hlx" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.950801 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d6648ca-f3df-4345-a2ac-d147e7ba85e1-operator-scripts\") pod \"glance-5610-account-create-update-gwfrn\" (UID: \"2d6648ca-f3df-4345-a2ac-d147e7ba85e1\") " pod="openstack/glance-5610-account-create-update-gwfrn" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.968721 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77sjh\" (UniqueName: \"kubernetes.io/projected/2d6648ca-f3df-4345-a2ac-d147e7ba85e1-kube-api-access-77sjh\") pod \"glance-5610-account-create-update-gwfrn\" (UID: \"2d6648ca-f3df-4345-a2ac-d147e7ba85e1\") " pod="openstack/glance-5610-account-create-update-gwfrn" Nov 29 04:54:20 crc kubenswrapper[4799]: I1129 04:54:20.968767 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92rbj\" (UniqueName: \"kubernetes.io/projected/f356f37d-81e1-4554-abfe-8d13412ae662-kube-api-access-92rbj\") pod \"glance-db-create-s2hlx\" (UID: \"f356f37d-81e1-4554-abfe-8d13412ae662\") " pod="openstack/glance-db-create-s2hlx" Nov 29 04:54:21 crc kubenswrapper[4799]: I1129 04:54:21.018404 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-s2hlx" Nov 29 04:54:21 crc kubenswrapper[4799]: I1129 04:54:21.035552 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-5610-account-create-update-gwfrn" Nov 29 04:54:21 crc kubenswrapper[4799]: I1129 04:54:21.481773 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-5610-account-create-update-gwfrn"] Nov 29 04:54:21 crc kubenswrapper[4799]: I1129 04:54:21.568222 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-s2hlx"] Nov 29 04:54:21 crc kubenswrapper[4799]: I1129 04:54:21.671037 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-5610-account-create-update-gwfrn" event={"ID":"2d6648ca-f3df-4345-a2ac-d147e7ba85e1","Type":"ContainerStarted","Data":"4110da79148696ba0b3fb864c1d2423183a041c611f45bcfda7bc86c475de7ac"} Nov 29 04:54:21 crc kubenswrapper[4799]: I1129 04:54:21.680166 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"c8c0babb-3163-4a72-a6d5-4fc5e64bee31","Type":"ContainerStarted","Data":"97c1f42d7b298cf31516f4a5972cb3412bc54950d3df2eb36b241c1ca591609f"} Nov 29 04:54:21 crc kubenswrapper[4799]: I1129 04:54:21.707121 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-s2hlx" event={"ID":"f356f37d-81e1-4554-abfe-8d13412ae662","Type":"ContainerStarted","Data":"8eaf06eb82ea8819c31801d3dbd9bd599bfa1b41788d3695fa6983d2b77ad1eb"} Nov 29 04:54:22 crc kubenswrapper[4799]: I1129 04:54:22.719428 4799 generic.go:334] "Generic (PLEG): container finished" podID="f356f37d-81e1-4554-abfe-8d13412ae662" containerID="34ce7b64a0457b12f9c8d2672d6716de8f796b0b907a0d2e1345c17138988137" exitCode=0 Nov 29 04:54:22 crc kubenswrapper[4799]: I1129 04:54:22.719496 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-s2hlx" event={"ID":"f356f37d-81e1-4554-abfe-8d13412ae662","Type":"ContainerDied","Data":"34ce7b64a0457b12f9c8d2672d6716de8f796b0b907a0d2e1345c17138988137"} Nov 29 04:54:22 crc kubenswrapper[4799]: I1129 04:54:22.722601 4799 generic.go:334] "Generic (PLEG): container finished" podID="2d6648ca-f3df-4345-a2ac-d147e7ba85e1" containerID="5c4d07e7972fb0dfc57af15e6f3f04e1dd0960514c48510b17ec18a5adb1a08c" exitCode=0 Nov 29 04:54:22 crc kubenswrapper[4799]: I1129 04:54:22.722668 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-5610-account-create-update-gwfrn" event={"ID":"2d6648ca-f3df-4345-a2ac-d147e7ba85e1","Type":"ContainerDied","Data":"5c4d07e7972fb0dfc57af15e6f3f04e1dd0960514c48510b17ec18a5adb1a08c"} Nov 29 04:54:23 crc kubenswrapper[4799]: I1129 04:54:23.228521 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-24kgw" Nov 29 04:54:23 crc kubenswrapper[4799]: I1129 04:54:23.228682 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-24kgw" Nov 29 04:54:23 crc kubenswrapper[4799]: I1129 04:54:23.279681 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-24kgw" Nov 29 04:54:23 crc kubenswrapper[4799]: I1129 04:54:23.736711 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"c8c0babb-3163-4a72-a6d5-4fc5e64bee31","Type":"ContainerStarted","Data":"78377ebd20b33b2bfa2f4119a2cb52f265368edba85acba7f76049059b4074dc"} Nov 29 04:54:23 crc kubenswrapper[4799]: I1129 04:54:23.736904 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"c8c0babb-3163-4a72-a6d5-4fc5e64bee31","Type":"ContainerStarted","Data":"df7ae523eb9f51570e760d55bdf08feeebf730bec6791577cce99fa3ab401ead"} Nov 29 04:54:23 crc kubenswrapper[4799]: I1129 04:54:23.736996 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Nov 29 04:54:23 crc kubenswrapper[4799]: I1129 04:54:23.778908 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.197083085 podStartE2EDuration="3.778862085s" podCreationTimestamp="2025-11-29 04:54:20 +0000 UTC" firstStartedPulling="2025-11-29 04:54:20.90219197 +0000 UTC m=+936.545122380" lastFinishedPulling="2025-11-29 04:54:22.48397098 +0000 UTC m=+938.126901380" observedRunningTime="2025-11-29 04:54:23.775026884 +0000 UTC m=+939.417957324" watchObservedRunningTime="2025-11-29 04:54:23.778862085 +0000 UTC m=+939.421792505" Nov 29 04:54:24 crc kubenswrapper[4799]: I1129 04:54:24.130815 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-5610-account-create-update-gwfrn" Nov 29 04:54:24 crc kubenswrapper[4799]: I1129 04:54:24.213460 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d6648ca-f3df-4345-a2ac-d147e7ba85e1-operator-scripts\") pod \"2d6648ca-f3df-4345-a2ac-d147e7ba85e1\" (UID: \"2d6648ca-f3df-4345-a2ac-d147e7ba85e1\") " Nov 29 04:54:24 crc kubenswrapper[4799]: I1129 04:54:24.213825 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-77sjh\" (UniqueName: \"kubernetes.io/projected/2d6648ca-f3df-4345-a2ac-d147e7ba85e1-kube-api-access-77sjh\") pod \"2d6648ca-f3df-4345-a2ac-d147e7ba85e1\" (UID: \"2d6648ca-f3df-4345-a2ac-d147e7ba85e1\") " Nov 29 04:54:24 crc kubenswrapper[4799]: I1129 04:54:24.214201 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d6648ca-f3df-4345-a2ac-d147e7ba85e1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2d6648ca-f3df-4345-a2ac-d147e7ba85e1" (UID: "2d6648ca-f3df-4345-a2ac-d147e7ba85e1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:54:24 crc kubenswrapper[4799]: I1129 04:54:24.214316 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d6648ca-f3df-4345-a2ac-d147e7ba85e1-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:24 crc kubenswrapper[4799]: I1129 04:54:24.221847 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d6648ca-f3df-4345-a2ac-d147e7ba85e1-kube-api-access-77sjh" (OuterVolumeSpecName: "kube-api-access-77sjh") pod "2d6648ca-f3df-4345-a2ac-d147e7ba85e1" (UID: "2d6648ca-f3df-4345-a2ac-d147e7ba85e1"). InnerVolumeSpecName "kube-api-access-77sjh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:54:24 crc kubenswrapper[4799]: I1129 04:54:24.279731 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-s2hlx" Nov 29 04:54:24 crc kubenswrapper[4799]: I1129 04:54:24.316458 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92rbj\" (UniqueName: \"kubernetes.io/projected/f356f37d-81e1-4554-abfe-8d13412ae662-kube-api-access-92rbj\") pod \"f356f37d-81e1-4554-abfe-8d13412ae662\" (UID: \"f356f37d-81e1-4554-abfe-8d13412ae662\") " Nov 29 04:54:24 crc kubenswrapper[4799]: I1129 04:54:24.316609 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f356f37d-81e1-4554-abfe-8d13412ae662-operator-scripts\") pod \"f356f37d-81e1-4554-abfe-8d13412ae662\" (UID: \"f356f37d-81e1-4554-abfe-8d13412ae662\") " Nov 29 04:54:24 crc kubenswrapper[4799]: I1129 04:54:24.317243 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f356f37d-81e1-4554-abfe-8d13412ae662-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f356f37d-81e1-4554-abfe-8d13412ae662" (UID: "f356f37d-81e1-4554-abfe-8d13412ae662"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:54:24 crc kubenswrapper[4799]: I1129 04:54:24.317300 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-77sjh\" (UniqueName: \"kubernetes.io/projected/2d6648ca-f3df-4345-a2ac-d147e7ba85e1-kube-api-access-77sjh\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:24 crc kubenswrapper[4799]: I1129 04:54:24.319359 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f356f37d-81e1-4554-abfe-8d13412ae662-kube-api-access-92rbj" (OuterVolumeSpecName: "kube-api-access-92rbj") pod "f356f37d-81e1-4554-abfe-8d13412ae662" (UID: "f356f37d-81e1-4554-abfe-8d13412ae662"). InnerVolumeSpecName "kube-api-access-92rbj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:54:24 crc kubenswrapper[4799]: I1129 04:54:24.422869 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f356f37d-81e1-4554-abfe-8d13412ae662-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:24 crc kubenswrapper[4799]: I1129 04:54:24.422918 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92rbj\" (UniqueName: \"kubernetes.io/projected/f356f37d-81e1-4554-abfe-8d13412ae662-kube-api-access-92rbj\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:24 crc kubenswrapper[4799]: I1129 04:54:24.752200 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-s2hlx" event={"ID":"f356f37d-81e1-4554-abfe-8d13412ae662","Type":"ContainerDied","Data":"8eaf06eb82ea8819c31801d3dbd9bd599bfa1b41788d3695fa6983d2b77ad1eb"} Nov 29 04:54:24 crc kubenswrapper[4799]: I1129 04:54:24.752280 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8eaf06eb82ea8819c31801d3dbd9bd599bfa1b41788d3695fa6983d2b77ad1eb" Nov 29 04:54:24 crc kubenswrapper[4799]: I1129 04:54:24.752568 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-s2hlx" Nov 29 04:54:24 crc kubenswrapper[4799]: I1129 04:54:24.753971 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-5610-account-create-update-gwfrn" event={"ID":"2d6648ca-f3df-4345-a2ac-d147e7ba85e1","Type":"ContainerDied","Data":"4110da79148696ba0b3fb864c1d2423183a041c611f45bcfda7bc86c475de7ac"} Nov 29 04:54:24 crc kubenswrapper[4799]: I1129 04:54:24.754013 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4110da79148696ba0b3fb864c1d2423183a041c611f45bcfda7bc86c475de7ac" Nov 29 04:54:24 crc kubenswrapper[4799]: I1129 04:54:24.754069 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-5610-account-create-update-gwfrn" Nov 29 04:54:24 crc kubenswrapper[4799]: I1129 04:54:24.974689 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-m5vgb"] Nov 29 04:54:24 crc kubenswrapper[4799]: E1129 04:54:24.975444 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f356f37d-81e1-4554-abfe-8d13412ae662" containerName="mariadb-database-create" Nov 29 04:54:24 crc kubenswrapper[4799]: I1129 04:54:24.975469 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f356f37d-81e1-4554-abfe-8d13412ae662" containerName="mariadb-database-create" Nov 29 04:54:24 crc kubenswrapper[4799]: E1129 04:54:24.975489 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d6648ca-f3df-4345-a2ac-d147e7ba85e1" containerName="mariadb-account-create-update" Nov 29 04:54:24 crc kubenswrapper[4799]: I1129 04:54:24.975496 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d6648ca-f3df-4345-a2ac-d147e7ba85e1" containerName="mariadb-account-create-update" Nov 29 04:54:24 crc kubenswrapper[4799]: I1129 04:54:24.975665 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d6648ca-f3df-4345-a2ac-d147e7ba85e1" containerName="mariadb-account-create-update" Nov 29 04:54:24 crc kubenswrapper[4799]: I1129 04:54:24.975711 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="f356f37d-81e1-4554-abfe-8d13412ae662" containerName="mariadb-database-create" Nov 29 04:54:24 crc kubenswrapper[4799]: I1129 04:54:24.976405 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-m5vgb" Nov 29 04:54:25 crc kubenswrapper[4799]: I1129 04:54:24.982578 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-m5vgb"] Nov 29 04:54:25 crc kubenswrapper[4799]: I1129 04:54:25.035728 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zlpw\" (UniqueName: \"kubernetes.io/projected/f67bb6d3-7729-467b-9bab-c3aba0a555ff-kube-api-access-4zlpw\") pod \"keystone-db-create-m5vgb\" (UID: \"f67bb6d3-7729-467b-9bab-c3aba0a555ff\") " pod="openstack/keystone-db-create-m5vgb" Nov 29 04:54:25 crc kubenswrapper[4799]: I1129 04:54:25.035817 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f67bb6d3-7729-467b-9bab-c3aba0a555ff-operator-scripts\") pod \"keystone-db-create-m5vgb\" (UID: \"f67bb6d3-7729-467b-9bab-c3aba0a555ff\") " pod="openstack/keystone-db-create-m5vgb" Nov 29 04:54:25 crc kubenswrapper[4799]: I1129 04:54:25.137823 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zlpw\" (UniqueName: \"kubernetes.io/projected/f67bb6d3-7729-467b-9bab-c3aba0a555ff-kube-api-access-4zlpw\") pod \"keystone-db-create-m5vgb\" (UID: \"f67bb6d3-7729-467b-9bab-c3aba0a555ff\") " pod="openstack/keystone-db-create-m5vgb" Nov 29 04:54:25 crc kubenswrapper[4799]: I1129 04:54:25.137912 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f67bb6d3-7729-467b-9bab-c3aba0a555ff-operator-scripts\") pod \"keystone-db-create-m5vgb\" (UID: \"f67bb6d3-7729-467b-9bab-c3aba0a555ff\") " pod="openstack/keystone-db-create-m5vgb" Nov 29 04:54:25 crc kubenswrapper[4799]: I1129 04:54:25.138908 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f67bb6d3-7729-467b-9bab-c3aba0a555ff-operator-scripts\") pod \"keystone-db-create-m5vgb\" (UID: \"f67bb6d3-7729-467b-9bab-c3aba0a555ff\") " pod="openstack/keystone-db-create-m5vgb" Nov 29 04:54:25 crc kubenswrapper[4799]: I1129 04:54:25.159827 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zlpw\" (UniqueName: \"kubernetes.io/projected/f67bb6d3-7729-467b-9bab-c3aba0a555ff-kube-api-access-4zlpw\") pod \"keystone-db-create-m5vgb\" (UID: \"f67bb6d3-7729-467b-9bab-c3aba0a555ff\") " pod="openstack/keystone-db-create-m5vgb" Nov 29 04:54:25 crc kubenswrapper[4799]: I1129 04:54:25.181037 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-53da-account-create-update-xrx67"] Nov 29 04:54:25 crc kubenswrapper[4799]: I1129 04:54:25.182171 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-53da-account-create-update-xrx67" Nov 29 04:54:25 crc kubenswrapper[4799]: I1129 04:54:25.187946 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Nov 29 04:54:25 crc kubenswrapper[4799]: I1129 04:54:25.229161 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-53da-account-create-update-xrx67"] Nov 29 04:54:25 crc kubenswrapper[4799]: I1129 04:54:25.239618 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/58450435-f88c-4ad2-abd9-ff0c1c183527-operator-scripts\") pod \"keystone-53da-account-create-update-xrx67\" (UID: \"58450435-f88c-4ad2-abd9-ff0c1c183527\") " pod="openstack/keystone-53da-account-create-update-xrx67" Nov 29 04:54:25 crc kubenswrapper[4799]: I1129 04:54:25.239705 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7b7g6\" (UniqueName: \"kubernetes.io/projected/58450435-f88c-4ad2-abd9-ff0c1c183527-kube-api-access-7b7g6\") pod \"keystone-53da-account-create-update-xrx67\" (UID: \"58450435-f88c-4ad2-abd9-ff0c1c183527\") " pod="openstack/keystone-53da-account-create-update-xrx67" Nov 29 04:54:25 crc kubenswrapper[4799]: I1129 04:54:25.318100 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-rsmmk"] Nov 29 04:54:25 crc kubenswrapper[4799]: I1129 04:54:25.320125 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-rsmmk" Nov 29 04:54:25 crc kubenswrapper[4799]: I1129 04:54:25.327321 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-m5vgb" Nov 29 04:54:25 crc kubenswrapper[4799]: I1129 04:54:25.329625 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-rsmmk"] Nov 29 04:54:25 crc kubenswrapper[4799]: I1129 04:54:25.341194 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/58450435-f88c-4ad2-abd9-ff0c1c183527-operator-scripts\") pod \"keystone-53da-account-create-update-xrx67\" (UID: \"58450435-f88c-4ad2-abd9-ff0c1c183527\") " pod="openstack/keystone-53da-account-create-update-xrx67" Nov 29 04:54:25 crc kubenswrapper[4799]: I1129 04:54:25.341277 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7b7g6\" (UniqueName: \"kubernetes.io/projected/58450435-f88c-4ad2-abd9-ff0c1c183527-kube-api-access-7b7g6\") pod \"keystone-53da-account-create-update-xrx67\" (UID: \"58450435-f88c-4ad2-abd9-ff0c1c183527\") " pod="openstack/keystone-53da-account-create-update-xrx67" Nov 29 04:54:25 crc kubenswrapper[4799]: I1129 04:54:25.346066 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/58450435-f88c-4ad2-abd9-ff0c1c183527-operator-scripts\") pod \"keystone-53da-account-create-update-xrx67\" (UID: \"58450435-f88c-4ad2-abd9-ff0c1c183527\") " pod="openstack/keystone-53da-account-create-update-xrx67" Nov 29 04:54:25 crc kubenswrapper[4799]: I1129 04:54:25.380001 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7b7g6\" (UniqueName: \"kubernetes.io/projected/58450435-f88c-4ad2-abd9-ff0c1c183527-kube-api-access-7b7g6\") pod \"keystone-53da-account-create-update-xrx67\" (UID: \"58450435-f88c-4ad2-abd9-ff0c1c183527\") " pod="openstack/keystone-53da-account-create-update-xrx67" Nov 29 04:54:25 crc kubenswrapper[4799]: I1129 04:54:25.388871 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-66ef-account-create-update-mcd96"] Nov 29 04:54:25 crc kubenswrapper[4799]: I1129 04:54:25.391090 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-66ef-account-create-update-mcd96" Nov 29 04:54:25 crc kubenswrapper[4799]: I1129 04:54:25.393957 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Nov 29 04:54:25 crc kubenswrapper[4799]: I1129 04:54:25.405544 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-66ef-account-create-update-mcd96"] Nov 29 04:54:25 crc kubenswrapper[4799]: I1129 04:54:25.441599 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jpq6g" Nov 29 04:54:25 crc kubenswrapper[4799]: I1129 04:54:25.441646 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jpq6g" Nov 29 04:54:25 crc kubenswrapper[4799]: I1129 04:54:25.442637 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9e8bbbf6-43e4-4cd2-a3cd-326d3d11009f-operator-scripts\") pod \"placement-db-create-rsmmk\" (UID: \"9e8bbbf6-43e4-4cd2-a3cd-326d3d11009f\") " pod="openstack/placement-db-create-rsmmk" Nov 29 04:54:25 crc kubenswrapper[4799]: I1129 04:54:25.442696 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlptw\" (UniqueName: \"kubernetes.io/projected/9e8bbbf6-43e4-4cd2-a3cd-326d3d11009f-kube-api-access-rlptw\") pod \"placement-db-create-rsmmk\" (UID: \"9e8bbbf6-43e4-4cd2-a3cd-326d3d11009f\") " pod="openstack/placement-db-create-rsmmk" Nov 29 04:54:25 crc kubenswrapper[4799]: I1129 04:54:25.442733 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmqvl\" (UniqueName: \"kubernetes.io/projected/74500694-328d-483c-8549-2bc90c72c811-kube-api-access-kmqvl\") pod \"placement-66ef-account-create-update-mcd96\" (UID: \"74500694-328d-483c-8549-2bc90c72c811\") " pod="openstack/placement-66ef-account-create-update-mcd96" Nov 29 04:54:25 crc kubenswrapper[4799]: I1129 04:54:25.442781 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/74500694-328d-483c-8549-2bc90c72c811-operator-scripts\") pod \"placement-66ef-account-create-update-mcd96\" (UID: \"74500694-328d-483c-8549-2bc90c72c811\") " pod="openstack/placement-66ef-account-create-update-mcd96" Nov 29 04:54:25 crc kubenswrapper[4799]: I1129 04:54:25.506084 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jpq6g" Nov 29 04:54:25 crc kubenswrapper[4799]: I1129 04:54:25.545243 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9e8bbbf6-43e4-4cd2-a3cd-326d3d11009f-operator-scripts\") pod \"placement-db-create-rsmmk\" (UID: \"9e8bbbf6-43e4-4cd2-a3cd-326d3d11009f\") " pod="openstack/placement-db-create-rsmmk" Nov 29 04:54:25 crc kubenswrapper[4799]: I1129 04:54:25.545314 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlptw\" (UniqueName: \"kubernetes.io/projected/9e8bbbf6-43e4-4cd2-a3cd-326d3d11009f-kube-api-access-rlptw\") pod \"placement-db-create-rsmmk\" (UID: \"9e8bbbf6-43e4-4cd2-a3cd-326d3d11009f\") " pod="openstack/placement-db-create-rsmmk" Nov 29 04:54:25 crc kubenswrapper[4799]: I1129 04:54:25.545362 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmqvl\" (UniqueName: \"kubernetes.io/projected/74500694-328d-483c-8549-2bc90c72c811-kube-api-access-kmqvl\") pod \"placement-66ef-account-create-update-mcd96\" (UID: \"74500694-328d-483c-8549-2bc90c72c811\") " pod="openstack/placement-66ef-account-create-update-mcd96" Nov 29 04:54:25 crc kubenswrapper[4799]: I1129 04:54:25.545422 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/74500694-328d-483c-8549-2bc90c72c811-operator-scripts\") pod \"placement-66ef-account-create-update-mcd96\" (UID: \"74500694-328d-483c-8549-2bc90c72c811\") " pod="openstack/placement-66ef-account-create-update-mcd96" Nov 29 04:54:25 crc kubenswrapper[4799]: I1129 04:54:25.546448 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/74500694-328d-483c-8549-2bc90c72c811-operator-scripts\") pod \"placement-66ef-account-create-update-mcd96\" (UID: \"74500694-328d-483c-8549-2bc90c72c811\") " pod="openstack/placement-66ef-account-create-update-mcd96" Nov 29 04:54:25 crc kubenswrapper[4799]: I1129 04:54:25.546902 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9e8bbbf6-43e4-4cd2-a3cd-326d3d11009f-operator-scripts\") pod \"placement-db-create-rsmmk\" (UID: \"9e8bbbf6-43e4-4cd2-a3cd-326d3d11009f\") " pod="openstack/placement-db-create-rsmmk" Nov 29 04:54:25 crc kubenswrapper[4799]: I1129 04:54:25.549421 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-53da-account-create-update-xrx67" Nov 29 04:54:25 crc kubenswrapper[4799]: I1129 04:54:25.573278 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlptw\" (UniqueName: \"kubernetes.io/projected/9e8bbbf6-43e4-4cd2-a3cd-326d3d11009f-kube-api-access-rlptw\") pod \"placement-db-create-rsmmk\" (UID: \"9e8bbbf6-43e4-4cd2-a3cd-326d3d11009f\") " pod="openstack/placement-db-create-rsmmk" Nov 29 04:54:25 crc kubenswrapper[4799]: I1129 04:54:25.578482 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmqvl\" (UniqueName: \"kubernetes.io/projected/74500694-328d-483c-8549-2bc90c72c811-kube-api-access-kmqvl\") pod \"placement-66ef-account-create-update-mcd96\" (UID: \"74500694-328d-483c-8549-2bc90c72c811\") " pod="openstack/placement-66ef-account-create-update-mcd96" Nov 29 04:54:25 crc kubenswrapper[4799]: I1129 04:54:25.646149 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-rsmmk" Nov 29 04:54:25 crc kubenswrapper[4799]: I1129 04:54:25.742541 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-66ef-account-create-update-mcd96" Nov 29 04:54:25 crc kubenswrapper[4799]: I1129 04:54:25.846645 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jpq6g" Nov 29 04:54:25 crc kubenswrapper[4799]: I1129 04:54:25.852428 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-m5vgb"] Nov 29 04:54:25 crc kubenswrapper[4799]: I1129 04:54:25.856679 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7fd796d7df-554d4" Nov 29 04:54:25 crc kubenswrapper[4799]: I1129 04:54:25.996917 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86db49b7ff-dbv6d" Nov 29 04:54:26 crc kubenswrapper[4799]: I1129 04:54:26.061318 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-88jbn"] Nov 29 04:54:26 crc kubenswrapper[4799]: I1129 04:54:26.065309 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-88jbn" Nov 29 04:54:26 crc kubenswrapper[4799]: I1129 04:54:26.069295 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Nov 29 04:54:26 crc kubenswrapper[4799]: I1129 04:54:26.070674 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-fhqdz" Nov 29 04:54:26 crc kubenswrapper[4799]: I1129 04:54:26.096431 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8sfw\" (UniqueName: \"kubernetes.io/projected/c77dc962-f9a7-40dc-aeb6-d07e9c1be588-kube-api-access-d8sfw\") pod \"glance-db-sync-88jbn\" (UID: \"c77dc962-f9a7-40dc-aeb6-d07e9c1be588\") " pod="openstack/glance-db-sync-88jbn" Nov 29 04:54:26 crc kubenswrapper[4799]: I1129 04:54:26.096906 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c77dc962-f9a7-40dc-aeb6-d07e9c1be588-db-sync-config-data\") pod \"glance-db-sync-88jbn\" (UID: \"c77dc962-f9a7-40dc-aeb6-d07e9c1be588\") " pod="openstack/glance-db-sync-88jbn" Nov 29 04:54:26 crc kubenswrapper[4799]: I1129 04:54:26.097059 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c77dc962-f9a7-40dc-aeb6-d07e9c1be588-config-data\") pod \"glance-db-sync-88jbn\" (UID: \"c77dc962-f9a7-40dc-aeb6-d07e9c1be588\") " pod="openstack/glance-db-sync-88jbn" Nov 29 04:54:26 crc kubenswrapper[4799]: I1129 04:54:26.097453 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c77dc962-f9a7-40dc-aeb6-d07e9c1be588-combined-ca-bundle\") pod \"glance-db-sync-88jbn\" (UID: \"c77dc962-f9a7-40dc-aeb6-d07e9c1be588\") " pod="openstack/glance-db-sync-88jbn" Nov 29 04:54:26 crc kubenswrapper[4799]: I1129 04:54:26.116978 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-88jbn"] Nov 29 04:54:26 crc kubenswrapper[4799]: I1129 04:54:26.148969 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-53da-account-create-update-xrx67"] Nov 29 04:54:26 crc kubenswrapper[4799]: I1129 04:54:26.200647 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c77dc962-f9a7-40dc-aeb6-d07e9c1be588-combined-ca-bundle\") pod \"glance-db-sync-88jbn\" (UID: \"c77dc962-f9a7-40dc-aeb6-d07e9c1be588\") " pod="openstack/glance-db-sync-88jbn" Nov 29 04:54:26 crc kubenswrapper[4799]: I1129 04:54:26.201011 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8sfw\" (UniqueName: \"kubernetes.io/projected/c77dc962-f9a7-40dc-aeb6-d07e9c1be588-kube-api-access-d8sfw\") pod \"glance-db-sync-88jbn\" (UID: \"c77dc962-f9a7-40dc-aeb6-d07e9c1be588\") " pod="openstack/glance-db-sync-88jbn" Nov 29 04:54:26 crc kubenswrapper[4799]: I1129 04:54:26.201157 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c77dc962-f9a7-40dc-aeb6-d07e9c1be588-db-sync-config-data\") pod \"glance-db-sync-88jbn\" (UID: \"c77dc962-f9a7-40dc-aeb6-d07e9c1be588\") " pod="openstack/glance-db-sync-88jbn" Nov 29 04:54:26 crc kubenswrapper[4799]: I1129 04:54:26.201245 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c77dc962-f9a7-40dc-aeb6-d07e9c1be588-config-data\") pod \"glance-db-sync-88jbn\" (UID: \"c77dc962-f9a7-40dc-aeb6-d07e9c1be588\") " pod="openstack/glance-db-sync-88jbn" Nov 29 04:54:26 crc kubenswrapper[4799]: I1129 04:54:26.222763 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c77dc962-f9a7-40dc-aeb6-d07e9c1be588-db-sync-config-data\") pod \"glance-db-sync-88jbn\" (UID: \"c77dc962-f9a7-40dc-aeb6-d07e9c1be588\") " pod="openstack/glance-db-sync-88jbn" Nov 29 04:54:26 crc kubenswrapper[4799]: I1129 04:54:26.233704 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c77dc962-f9a7-40dc-aeb6-d07e9c1be588-combined-ca-bundle\") pod \"glance-db-sync-88jbn\" (UID: \"c77dc962-f9a7-40dc-aeb6-d07e9c1be588\") " pod="openstack/glance-db-sync-88jbn" Nov 29 04:54:26 crc kubenswrapper[4799]: I1129 04:54:26.233751 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c77dc962-f9a7-40dc-aeb6-d07e9c1be588-config-data\") pod \"glance-db-sync-88jbn\" (UID: \"c77dc962-f9a7-40dc-aeb6-d07e9c1be588\") " pod="openstack/glance-db-sync-88jbn" Nov 29 04:54:26 crc kubenswrapper[4799]: I1129 04:54:26.250100 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jpq6g"] Nov 29 04:54:26 crc kubenswrapper[4799]: I1129 04:54:26.278690 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8sfw\" (UniqueName: \"kubernetes.io/projected/c77dc962-f9a7-40dc-aeb6-d07e9c1be588-kube-api-access-d8sfw\") pod \"glance-db-sync-88jbn\" (UID: \"c77dc962-f9a7-40dc-aeb6-d07e9c1be588\") " pod="openstack/glance-db-sync-88jbn" Nov 29 04:54:26 crc kubenswrapper[4799]: I1129 04:54:26.313651 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-554d4"] Nov 29 04:54:26 crc kubenswrapper[4799]: I1129 04:54:26.374654 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-rsmmk"] Nov 29 04:54:26 crc kubenswrapper[4799]: I1129 04:54:26.472215 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-88jbn" Nov 29 04:54:26 crc kubenswrapper[4799]: I1129 04:54:26.476652 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-66ef-account-create-update-mcd96"] Nov 29 04:54:26 crc kubenswrapper[4799]: I1129 04:54:26.581628 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-kkdw8" podUID="1d3d81e5-591a-469d-9851-86271162d455" containerName="ovn-controller" probeResult="failure" output=< Nov 29 04:54:26 crc kubenswrapper[4799]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 29 04:54:26 crc kubenswrapper[4799]: > Nov 29 04:54:26 crc kubenswrapper[4799]: I1129 04:54:26.786679 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-66ef-account-create-update-mcd96" event={"ID":"74500694-328d-483c-8549-2bc90c72c811","Type":"ContainerStarted","Data":"5546b37350112c0c7c5948eae46937acf3e343ca5f0ea9f84e51111a934a2612"} Nov 29 04:54:26 crc kubenswrapper[4799]: I1129 04:54:26.789735 4799 generic.go:334] "Generic (PLEG): container finished" podID="f67bb6d3-7729-467b-9bab-c3aba0a555ff" containerID="4fdff7a088d727ee5ebb687fd9227d1917f4ab6b2bc9a6c9827f0958c39caad0" exitCode=0 Nov 29 04:54:26 crc kubenswrapper[4799]: I1129 04:54:26.789860 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-m5vgb" event={"ID":"f67bb6d3-7729-467b-9bab-c3aba0a555ff","Type":"ContainerDied","Data":"4fdff7a088d727ee5ebb687fd9227d1917f4ab6b2bc9a6c9827f0958c39caad0"} Nov 29 04:54:26 crc kubenswrapper[4799]: I1129 04:54:26.789879 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-m5vgb" event={"ID":"f67bb6d3-7729-467b-9bab-c3aba0a555ff","Type":"ContainerStarted","Data":"aaac4ad3ad22bacbf771fbcb044efcb848046a4ae942267e114213e17ed5af7f"} Nov 29 04:54:26 crc kubenswrapper[4799]: I1129 04:54:26.792252 4799 generic.go:334] "Generic (PLEG): container finished" podID="58450435-f88c-4ad2-abd9-ff0c1c183527" containerID="15012f9018ab27ea86dbb3801f96976cbb660267030da824f8195264a2330db5" exitCode=0 Nov 29 04:54:26 crc kubenswrapper[4799]: I1129 04:54:26.792377 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-53da-account-create-update-xrx67" event={"ID":"58450435-f88c-4ad2-abd9-ff0c1c183527","Type":"ContainerDied","Data":"15012f9018ab27ea86dbb3801f96976cbb660267030da824f8195264a2330db5"} Nov 29 04:54:26 crc kubenswrapper[4799]: I1129 04:54:26.792467 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-53da-account-create-update-xrx67" event={"ID":"58450435-f88c-4ad2-abd9-ff0c1c183527","Type":"ContainerStarted","Data":"449c8fcaa928fb79365b7d9b70e9f54507aebb3f95cf5737a2c5c7d53ec772d0"} Nov 29 04:54:26 crc kubenswrapper[4799]: I1129 04:54:26.801385 4799 generic.go:334] "Generic (PLEG): container finished" podID="9e8bbbf6-43e4-4cd2-a3cd-326d3d11009f" containerID="ce9bfcee7e64c14e78367370775cfb0d18567f3e71859d42cd40a5d8bb798705" exitCode=0 Nov 29 04:54:26 crc kubenswrapper[4799]: I1129 04:54:26.801437 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-rsmmk" event={"ID":"9e8bbbf6-43e4-4cd2-a3cd-326d3d11009f","Type":"ContainerDied","Data":"ce9bfcee7e64c14e78367370775cfb0d18567f3e71859d42cd40a5d8bb798705"} Nov 29 04:54:26 crc kubenswrapper[4799]: I1129 04:54:26.801493 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-rsmmk" event={"ID":"9e8bbbf6-43e4-4cd2-a3cd-326d3d11009f","Type":"ContainerStarted","Data":"ca9a62687c5f00e4d0910d51413d0b8a088a40778aeb9e5e31fd936a4db75aee"} Nov 29 04:54:26 crc kubenswrapper[4799]: I1129 04:54:26.801661 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7fd796d7df-554d4" podUID="6b3734a5-609e-4ffc-ac55-1d9fb1e699cc" containerName="dnsmasq-dns" containerID="cri-o://662675aecdd4c6f6a80b63ef0c29e8aaa6b5b4687e8314e732545cbd84a8e52f" gracePeriod=10 Nov 29 04:54:27 crc kubenswrapper[4799]: I1129 04:54:27.078306 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-88jbn"] Nov 29 04:54:27 crc kubenswrapper[4799]: I1129 04:54:27.290423 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-554d4" Nov 29 04:54:27 crc kubenswrapper[4799]: I1129 04:54:27.437217 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b3734a5-609e-4ffc-ac55-1d9fb1e699cc-dns-svc\") pod \"6b3734a5-609e-4ffc-ac55-1d9fb1e699cc\" (UID: \"6b3734a5-609e-4ffc-ac55-1d9fb1e699cc\") " Nov 29 04:54:27 crc kubenswrapper[4799]: I1129 04:54:27.437328 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b3734a5-609e-4ffc-ac55-1d9fb1e699cc-ovsdbserver-nb\") pod \"6b3734a5-609e-4ffc-ac55-1d9fb1e699cc\" (UID: \"6b3734a5-609e-4ffc-ac55-1d9fb1e699cc\") " Nov 29 04:54:27 crc kubenswrapper[4799]: I1129 04:54:27.437508 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b3734a5-609e-4ffc-ac55-1d9fb1e699cc-config\") pod \"6b3734a5-609e-4ffc-ac55-1d9fb1e699cc\" (UID: \"6b3734a5-609e-4ffc-ac55-1d9fb1e699cc\") " Nov 29 04:54:27 crc kubenswrapper[4799]: I1129 04:54:27.437569 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l96rk\" (UniqueName: \"kubernetes.io/projected/6b3734a5-609e-4ffc-ac55-1d9fb1e699cc-kube-api-access-l96rk\") pod \"6b3734a5-609e-4ffc-ac55-1d9fb1e699cc\" (UID: \"6b3734a5-609e-4ffc-ac55-1d9fb1e699cc\") " Nov 29 04:54:27 crc kubenswrapper[4799]: I1129 04:54:27.444656 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b3734a5-609e-4ffc-ac55-1d9fb1e699cc-kube-api-access-l96rk" (OuterVolumeSpecName: "kube-api-access-l96rk") pod "6b3734a5-609e-4ffc-ac55-1d9fb1e699cc" (UID: "6b3734a5-609e-4ffc-ac55-1d9fb1e699cc"). InnerVolumeSpecName "kube-api-access-l96rk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:54:27 crc kubenswrapper[4799]: I1129 04:54:27.477759 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b3734a5-609e-4ffc-ac55-1d9fb1e699cc-config" (OuterVolumeSpecName: "config") pod "6b3734a5-609e-4ffc-ac55-1d9fb1e699cc" (UID: "6b3734a5-609e-4ffc-ac55-1d9fb1e699cc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:54:27 crc kubenswrapper[4799]: I1129 04:54:27.478411 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b3734a5-609e-4ffc-ac55-1d9fb1e699cc-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6b3734a5-609e-4ffc-ac55-1d9fb1e699cc" (UID: "6b3734a5-609e-4ffc-ac55-1d9fb1e699cc"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:54:27 crc kubenswrapper[4799]: I1129 04:54:27.480005 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b3734a5-609e-4ffc-ac55-1d9fb1e699cc-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6b3734a5-609e-4ffc-ac55-1d9fb1e699cc" (UID: "6b3734a5-609e-4ffc-ac55-1d9fb1e699cc"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:54:27 crc kubenswrapper[4799]: I1129 04:54:27.539726 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l96rk\" (UniqueName: \"kubernetes.io/projected/6b3734a5-609e-4ffc-ac55-1d9fb1e699cc-kube-api-access-l96rk\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:27 crc kubenswrapper[4799]: I1129 04:54:27.540531 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b3734a5-609e-4ffc-ac55-1d9fb1e699cc-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:27 crc kubenswrapper[4799]: I1129 04:54:27.540608 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b3734a5-609e-4ffc-ac55-1d9fb1e699cc-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:27 crc kubenswrapper[4799]: I1129 04:54:27.540706 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b3734a5-609e-4ffc-ac55-1d9fb1e699cc-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:27 crc kubenswrapper[4799]: I1129 04:54:27.812602 4799 generic.go:334] "Generic (PLEG): container finished" podID="74500694-328d-483c-8549-2bc90c72c811" containerID="12b12b9deb3bd8c8432d4bd911a69d63586a1503a1208c5f786aceee458ac305" exitCode=0 Nov 29 04:54:27 crc kubenswrapper[4799]: I1129 04:54:27.813171 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-66ef-account-create-update-mcd96" event={"ID":"74500694-328d-483c-8549-2bc90c72c811","Type":"ContainerDied","Data":"12b12b9deb3bd8c8432d4bd911a69d63586a1503a1208c5f786aceee458ac305"} Nov 29 04:54:27 crc kubenswrapper[4799]: I1129 04:54:27.814530 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-88jbn" event={"ID":"c77dc962-f9a7-40dc-aeb6-d07e9c1be588","Type":"ContainerStarted","Data":"10523f9715769be575965752a09ec60026d7d25d5a259f7484c0d2a5b591c4e6"} Nov 29 04:54:27 crc kubenswrapper[4799]: I1129 04:54:27.817672 4799 generic.go:334] "Generic (PLEG): container finished" podID="6b3734a5-609e-4ffc-ac55-1d9fb1e699cc" containerID="662675aecdd4c6f6a80b63ef0c29e8aaa6b5b4687e8314e732545cbd84a8e52f" exitCode=0 Nov 29 04:54:27 crc kubenswrapper[4799]: I1129 04:54:27.817721 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-554d4" event={"ID":"6b3734a5-609e-4ffc-ac55-1d9fb1e699cc","Type":"ContainerDied","Data":"662675aecdd4c6f6a80b63ef0c29e8aaa6b5b4687e8314e732545cbd84a8e52f"} Nov 29 04:54:27 crc kubenswrapper[4799]: I1129 04:54:27.817750 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-554d4" Nov 29 04:54:27 crc kubenswrapper[4799]: I1129 04:54:27.817774 4799 scope.go:117] "RemoveContainer" containerID="662675aecdd4c6f6a80b63ef0c29e8aaa6b5b4687e8314e732545cbd84a8e52f" Nov 29 04:54:27 crc kubenswrapper[4799]: I1129 04:54:27.817757 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-554d4" event={"ID":"6b3734a5-609e-4ffc-ac55-1d9fb1e699cc","Type":"ContainerDied","Data":"aac3f393379d0d3c6033bba5698a20279fbce5d538785dac9c57c7e12b1fcb6c"} Nov 29 04:54:27 crc kubenswrapper[4799]: I1129 04:54:27.818413 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jpq6g" podUID="89b0a912-606b-4d5d-bd40-09f89f2db649" containerName="registry-server" containerID="cri-o://7c548c9a584c5577fe8115c5f60db7f80c61da43605a9ba6dc27fd1fe751b204" gracePeriod=2 Nov 29 04:54:27 crc kubenswrapper[4799]: I1129 04:54:27.865641 4799 scope.go:117] "RemoveContainer" containerID="e91847d686e65cefcb856dfa8a8fd1a0b792ecec149e99d15246b3fc74fd2a63" Nov 29 04:54:27 crc kubenswrapper[4799]: I1129 04:54:27.875965 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-554d4"] Nov 29 04:54:27 crc kubenswrapper[4799]: I1129 04:54:27.884077 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-554d4"] Nov 29 04:54:27 crc kubenswrapper[4799]: I1129 04:54:27.956244 4799 scope.go:117] "RemoveContainer" containerID="662675aecdd4c6f6a80b63ef0c29e8aaa6b5b4687e8314e732545cbd84a8e52f" Nov 29 04:54:27 crc kubenswrapper[4799]: E1129 04:54:27.958067 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"662675aecdd4c6f6a80b63ef0c29e8aaa6b5b4687e8314e732545cbd84a8e52f\": container with ID starting with 662675aecdd4c6f6a80b63ef0c29e8aaa6b5b4687e8314e732545cbd84a8e52f not found: ID does not exist" containerID="662675aecdd4c6f6a80b63ef0c29e8aaa6b5b4687e8314e732545cbd84a8e52f" Nov 29 04:54:27 crc kubenswrapper[4799]: I1129 04:54:27.958135 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"662675aecdd4c6f6a80b63ef0c29e8aaa6b5b4687e8314e732545cbd84a8e52f"} err="failed to get container status \"662675aecdd4c6f6a80b63ef0c29e8aaa6b5b4687e8314e732545cbd84a8e52f\": rpc error: code = NotFound desc = could not find container \"662675aecdd4c6f6a80b63ef0c29e8aaa6b5b4687e8314e732545cbd84a8e52f\": container with ID starting with 662675aecdd4c6f6a80b63ef0c29e8aaa6b5b4687e8314e732545cbd84a8e52f not found: ID does not exist" Nov 29 04:54:27 crc kubenswrapper[4799]: I1129 04:54:27.958162 4799 scope.go:117] "RemoveContainer" containerID="e91847d686e65cefcb856dfa8a8fd1a0b792ecec149e99d15246b3fc74fd2a63" Nov 29 04:54:27 crc kubenswrapper[4799]: E1129 04:54:27.958560 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e91847d686e65cefcb856dfa8a8fd1a0b792ecec149e99d15246b3fc74fd2a63\": container with ID starting with e91847d686e65cefcb856dfa8a8fd1a0b792ecec149e99d15246b3fc74fd2a63 not found: ID does not exist" containerID="e91847d686e65cefcb856dfa8a8fd1a0b792ecec149e99d15246b3fc74fd2a63" Nov 29 04:54:27 crc kubenswrapper[4799]: I1129 04:54:27.958579 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e91847d686e65cefcb856dfa8a8fd1a0b792ecec149e99d15246b3fc74fd2a63"} err="failed to get container status \"e91847d686e65cefcb856dfa8a8fd1a0b792ecec149e99d15246b3fc74fd2a63\": rpc error: code = NotFound desc = could not find container \"e91847d686e65cefcb856dfa8a8fd1a0b792ecec149e99d15246b3fc74fd2a63\": container with ID starting with e91847d686e65cefcb856dfa8a8fd1a0b792ecec149e99d15246b3fc74fd2a63 not found: ID does not exist" Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.238713 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-53da-account-create-update-xrx67" Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.266566 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/58450435-f88c-4ad2-abd9-ff0c1c183527-operator-scripts\") pod \"58450435-f88c-4ad2-abd9-ff0c1c183527\" (UID: \"58450435-f88c-4ad2-abd9-ff0c1c183527\") " Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.267017 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7b7g6\" (UniqueName: \"kubernetes.io/projected/58450435-f88c-4ad2-abd9-ff0c1c183527-kube-api-access-7b7g6\") pod \"58450435-f88c-4ad2-abd9-ff0c1c183527\" (UID: \"58450435-f88c-4ad2-abd9-ff0c1c183527\") " Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.269533 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58450435-f88c-4ad2-abd9-ff0c1c183527-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "58450435-f88c-4ad2-abd9-ff0c1c183527" (UID: "58450435-f88c-4ad2-abd9-ff0c1c183527"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.275321 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58450435-f88c-4ad2-abd9-ff0c1c183527-kube-api-access-7b7g6" (OuterVolumeSpecName: "kube-api-access-7b7g6") pod "58450435-f88c-4ad2-abd9-ff0c1c183527" (UID: "58450435-f88c-4ad2-abd9-ff0c1c183527"). InnerVolumeSpecName "kube-api-access-7b7g6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.292707 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-rsmmk" Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.304642 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-m5vgb" Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.382257 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9e8bbbf6-43e4-4cd2-a3cd-326d3d11009f-operator-scripts\") pod \"9e8bbbf6-43e4-4cd2-a3cd-326d3d11009f\" (UID: \"9e8bbbf6-43e4-4cd2-a3cd-326d3d11009f\") " Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.382352 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zlpw\" (UniqueName: \"kubernetes.io/projected/f67bb6d3-7729-467b-9bab-c3aba0a555ff-kube-api-access-4zlpw\") pod \"f67bb6d3-7729-467b-9bab-c3aba0a555ff\" (UID: \"f67bb6d3-7729-467b-9bab-c3aba0a555ff\") " Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.382538 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rlptw\" (UniqueName: \"kubernetes.io/projected/9e8bbbf6-43e4-4cd2-a3cd-326d3d11009f-kube-api-access-rlptw\") pod \"9e8bbbf6-43e4-4cd2-a3cd-326d3d11009f\" (UID: \"9e8bbbf6-43e4-4cd2-a3cd-326d3d11009f\") " Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.382581 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f67bb6d3-7729-467b-9bab-c3aba0a555ff-operator-scripts\") pod \"f67bb6d3-7729-467b-9bab-c3aba0a555ff\" (UID: \"f67bb6d3-7729-467b-9bab-c3aba0a555ff\") " Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.385896 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7b7g6\" (UniqueName: \"kubernetes.io/projected/58450435-f88c-4ad2-abd9-ff0c1c183527-kube-api-access-7b7g6\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.385925 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/58450435-f88c-4ad2-abd9-ff0c1c183527-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.386734 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f67bb6d3-7729-467b-9bab-c3aba0a555ff-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f67bb6d3-7729-467b-9bab-c3aba0a555ff" (UID: "f67bb6d3-7729-467b-9bab-c3aba0a555ff"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.386865 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e8bbbf6-43e4-4cd2-a3cd-326d3d11009f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9e8bbbf6-43e4-4cd2-a3cd-326d3d11009f" (UID: "9e8bbbf6-43e4-4cd2-a3cd-326d3d11009f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.390051 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f67bb6d3-7729-467b-9bab-c3aba0a555ff-kube-api-access-4zlpw" (OuterVolumeSpecName: "kube-api-access-4zlpw") pod "f67bb6d3-7729-467b-9bab-c3aba0a555ff" (UID: "f67bb6d3-7729-467b-9bab-c3aba0a555ff"). InnerVolumeSpecName "kube-api-access-4zlpw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.390956 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e8bbbf6-43e4-4cd2-a3cd-326d3d11009f-kube-api-access-rlptw" (OuterVolumeSpecName: "kube-api-access-rlptw") pod "9e8bbbf6-43e4-4cd2-a3cd-326d3d11009f" (UID: "9e8bbbf6-43e4-4cd2-a3cd-326d3d11009f"). InnerVolumeSpecName "kube-api-access-rlptw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.404332 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jpq6g" Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.486763 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89b0a912-606b-4d5d-bd40-09f89f2db649-catalog-content\") pod \"89b0a912-606b-4d5d-bd40-09f89f2db649\" (UID: \"89b0a912-606b-4d5d-bd40-09f89f2db649\") " Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.486983 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89b0a912-606b-4d5d-bd40-09f89f2db649-utilities\") pod \"89b0a912-606b-4d5d-bd40-09f89f2db649\" (UID: \"89b0a912-606b-4d5d-bd40-09f89f2db649\") " Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.487049 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jfl99\" (UniqueName: \"kubernetes.io/projected/89b0a912-606b-4d5d-bd40-09f89f2db649-kube-api-access-jfl99\") pod \"89b0a912-606b-4d5d-bd40-09f89f2db649\" (UID: \"89b0a912-606b-4d5d-bd40-09f89f2db649\") " Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.487376 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9e8bbbf6-43e4-4cd2-a3cd-326d3d11009f-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.487396 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zlpw\" (UniqueName: \"kubernetes.io/projected/f67bb6d3-7729-467b-9bab-c3aba0a555ff-kube-api-access-4zlpw\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.487409 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rlptw\" (UniqueName: \"kubernetes.io/projected/9e8bbbf6-43e4-4cd2-a3cd-326d3d11009f-kube-api-access-rlptw\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.487421 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f67bb6d3-7729-467b-9bab-c3aba0a555ff-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.488320 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89b0a912-606b-4d5d-bd40-09f89f2db649-utilities" (OuterVolumeSpecName: "utilities") pod "89b0a912-606b-4d5d-bd40-09f89f2db649" (UID: "89b0a912-606b-4d5d-bd40-09f89f2db649"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.491324 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89b0a912-606b-4d5d-bd40-09f89f2db649-kube-api-access-jfl99" (OuterVolumeSpecName: "kube-api-access-jfl99") pod "89b0a912-606b-4d5d-bd40-09f89f2db649" (UID: "89b0a912-606b-4d5d-bd40-09f89f2db649"). InnerVolumeSpecName "kube-api-access-jfl99". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.590015 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89b0a912-606b-4d5d-bd40-09f89f2db649-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.590060 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jfl99\" (UniqueName: \"kubernetes.io/projected/89b0a912-606b-4d5d-bd40-09f89f2db649-kube-api-access-jfl99\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.594187 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89b0a912-606b-4d5d-bd40-09f89f2db649-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "89b0a912-606b-4d5d-bd40-09f89f2db649" (UID: "89b0a912-606b-4d5d-bd40-09f89f2db649"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.672825 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b3734a5-609e-4ffc-ac55-1d9fb1e699cc" path="/var/lib/kubelet/pods/6b3734a5-609e-4ffc-ac55-1d9fb1e699cc/volumes" Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.691338 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89b0a912-606b-4d5d-bd40-09f89f2db649-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.828518 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-53da-account-create-update-xrx67" event={"ID":"58450435-f88c-4ad2-abd9-ff0c1c183527","Type":"ContainerDied","Data":"449c8fcaa928fb79365b7d9b70e9f54507aebb3f95cf5737a2c5c7d53ec772d0"} Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.828597 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="449c8fcaa928fb79365b7d9b70e9f54507aebb3f95cf5737a2c5c7d53ec772d0" Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.828693 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-53da-account-create-update-xrx67" Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.831377 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-rsmmk" event={"ID":"9e8bbbf6-43e4-4cd2-a3cd-326d3d11009f","Type":"ContainerDied","Data":"ca9a62687c5f00e4d0910d51413d0b8a088a40778aeb9e5e31fd936a4db75aee"} Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.831445 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ca9a62687c5f00e4d0910d51413d0b8a088a40778aeb9e5e31fd936a4db75aee" Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.831403 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-rsmmk" Nov 29 04:54:28 crc kubenswrapper[4799]: E1129 04:54:28.837704 4799 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9e8bbbf6_43e4_4cd2_a3cd_326d3d11009f.slice/crio-ca9a62687c5f00e4d0910d51413d0b8a088a40778aeb9e5e31fd936a4db75aee\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf67bb6d3_7729_467b_9bab_c3aba0a555ff.slice/crio-aaac4ad3ad22bacbf771fbcb044efcb848046a4ae942267e114213e17ed5af7f\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf67bb6d3_7729_467b_9bab_c3aba0a555ff.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9e8bbbf6_43e4_4cd2_a3cd_326d3d11009f.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58450435_f88c_4ad2_abd9_ff0c1c183527.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod89b0a912_606b_4d5d_bd40_09f89f2db649.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1788e4d3_8c2a_4d06_b746_c0fb12514148.slice/crio-c7803748c9deb308ab6a3661ce9b6ea289206d8f547aa0262d26c7a7a46628cc.scope\": RecentStats: unable to find data in memory cache]" Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.838117 4799 generic.go:334] "Generic (PLEG): container finished" podID="89b0a912-606b-4d5d-bd40-09f89f2db649" containerID="7c548c9a584c5577fe8115c5f60db7f80c61da43605a9ba6dc27fd1fe751b204" exitCode=0 Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.838236 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jpq6g" Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.838916 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jpq6g" event={"ID":"89b0a912-606b-4d5d-bd40-09f89f2db649","Type":"ContainerDied","Data":"7c548c9a584c5577fe8115c5f60db7f80c61da43605a9ba6dc27fd1fe751b204"} Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.838980 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jpq6g" event={"ID":"89b0a912-606b-4d5d-bd40-09f89f2db649","Type":"ContainerDied","Data":"3b1bbf0db10e104f84595b8b29c5cddadb967e17cfa5e7c40da5f7326e0850ce"} Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.839015 4799 scope.go:117] "RemoveContainer" containerID="7c548c9a584c5577fe8115c5f60db7f80c61da43605a9ba6dc27fd1fe751b204" Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.844237 4799 generic.go:334] "Generic (PLEG): container finished" podID="1788e4d3-8c2a-4d06-b746-c0fb12514148" containerID="c7803748c9deb308ab6a3661ce9b6ea289206d8f547aa0262d26c7a7a46628cc" exitCode=0 Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.844403 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1788e4d3-8c2a-4d06-b746-c0fb12514148","Type":"ContainerDied","Data":"c7803748c9deb308ab6a3661ce9b6ea289206d8f547aa0262d26c7a7a46628cc"} Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.848134 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-m5vgb" event={"ID":"f67bb6d3-7729-467b-9bab-c3aba0a555ff","Type":"ContainerDied","Data":"aaac4ad3ad22bacbf771fbcb044efcb848046a4ae942267e114213e17ed5af7f"} Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.848169 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-m5vgb" Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.848190 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aaac4ad3ad22bacbf771fbcb044efcb848046a4ae942267e114213e17ed5af7f" Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.875318 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jpq6g"] Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.886301 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jpq6g"] Nov 29 04:54:28 crc kubenswrapper[4799]: I1129 04:54:28.890931 4799 scope.go:117] "RemoveContainer" containerID="54931734410bb6d7974694eac8365871e27917bb9d9e0801647e259119412efe" Nov 29 04:54:29 crc kubenswrapper[4799]: I1129 04:54:29.098346 4799 scope.go:117] "RemoveContainer" containerID="782739f273c3e17b3a2fe4efbe5a5517a28b476625a999543621bd4b130b3ffb" Nov 29 04:54:29 crc kubenswrapper[4799]: I1129 04:54:29.119393 4799 scope.go:117] "RemoveContainer" containerID="7c548c9a584c5577fe8115c5f60db7f80c61da43605a9ba6dc27fd1fe751b204" Nov 29 04:54:29 crc kubenswrapper[4799]: E1129 04:54:29.120205 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c548c9a584c5577fe8115c5f60db7f80c61da43605a9ba6dc27fd1fe751b204\": container with ID starting with 7c548c9a584c5577fe8115c5f60db7f80c61da43605a9ba6dc27fd1fe751b204 not found: ID does not exist" containerID="7c548c9a584c5577fe8115c5f60db7f80c61da43605a9ba6dc27fd1fe751b204" Nov 29 04:54:29 crc kubenswrapper[4799]: I1129 04:54:29.120262 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c548c9a584c5577fe8115c5f60db7f80c61da43605a9ba6dc27fd1fe751b204"} err="failed to get container status \"7c548c9a584c5577fe8115c5f60db7f80c61da43605a9ba6dc27fd1fe751b204\": rpc error: code = NotFound desc = could not find container \"7c548c9a584c5577fe8115c5f60db7f80c61da43605a9ba6dc27fd1fe751b204\": container with ID starting with 7c548c9a584c5577fe8115c5f60db7f80c61da43605a9ba6dc27fd1fe751b204 not found: ID does not exist" Nov 29 04:54:29 crc kubenswrapper[4799]: I1129 04:54:29.120295 4799 scope.go:117] "RemoveContainer" containerID="54931734410bb6d7974694eac8365871e27917bb9d9e0801647e259119412efe" Nov 29 04:54:29 crc kubenswrapper[4799]: E1129 04:54:29.120717 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54931734410bb6d7974694eac8365871e27917bb9d9e0801647e259119412efe\": container with ID starting with 54931734410bb6d7974694eac8365871e27917bb9d9e0801647e259119412efe not found: ID does not exist" containerID="54931734410bb6d7974694eac8365871e27917bb9d9e0801647e259119412efe" Nov 29 04:54:29 crc kubenswrapper[4799]: I1129 04:54:29.120756 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54931734410bb6d7974694eac8365871e27917bb9d9e0801647e259119412efe"} err="failed to get container status \"54931734410bb6d7974694eac8365871e27917bb9d9e0801647e259119412efe\": rpc error: code = NotFound desc = could not find container \"54931734410bb6d7974694eac8365871e27917bb9d9e0801647e259119412efe\": container with ID starting with 54931734410bb6d7974694eac8365871e27917bb9d9e0801647e259119412efe not found: ID does not exist" Nov 29 04:54:29 crc kubenswrapper[4799]: I1129 04:54:29.120783 4799 scope.go:117] "RemoveContainer" containerID="782739f273c3e17b3a2fe4efbe5a5517a28b476625a999543621bd4b130b3ffb" Nov 29 04:54:29 crc kubenswrapper[4799]: E1129 04:54:29.121219 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"782739f273c3e17b3a2fe4efbe5a5517a28b476625a999543621bd4b130b3ffb\": container with ID starting with 782739f273c3e17b3a2fe4efbe5a5517a28b476625a999543621bd4b130b3ffb not found: ID does not exist" containerID="782739f273c3e17b3a2fe4efbe5a5517a28b476625a999543621bd4b130b3ffb" Nov 29 04:54:29 crc kubenswrapper[4799]: I1129 04:54:29.121246 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"782739f273c3e17b3a2fe4efbe5a5517a28b476625a999543621bd4b130b3ffb"} err="failed to get container status \"782739f273c3e17b3a2fe4efbe5a5517a28b476625a999543621bd4b130b3ffb\": rpc error: code = NotFound desc = could not find container \"782739f273c3e17b3a2fe4efbe5a5517a28b476625a999543621bd4b130b3ffb\": container with ID starting with 782739f273c3e17b3a2fe4efbe5a5517a28b476625a999543621bd4b130b3ffb not found: ID does not exist" Nov 29 04:54:29 crc kubenswrapper[4799]: I1129 04:54:29.149707 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-66ef-account-create-update-mcd96" Nov 29 04:54:29 crc kubenswrapper[4799]: I1129 04:54:29.305328 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/74500694-328d-483c-8549-2bc90c72c811-operator-scripts\") pod \"74500694-328d-483c-8549-2bc90c72c811\" (UID: \"74500694-328d-483c-8549-2bc90c72c811\") " Nov 29 04:54:29 crc kubenswrapper[4799]: I1129 04:54:29.305526 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kmqvl\" (UniqueName: \"kubernetes.io/projected/74500694-328d-483c-8549-2bc90c72c811-kube-api-access-kmqvl\") pod \"74500694-328d-483c-8549-2bc90c72c811\" (UID: \"74500694-328d-483c-8549-2bc90c72c811\") " Nov 29 04:54:29 crc kubenswrapper[4799]: I1129 04:54:29.306549 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74500694-328d-483c-8549-2bc90c72c811-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "74500694-328d-483c-8549-2bc90c72c811" (UID: "74500694-328d-483c-8549-2bc90c72c811"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:54:29 crc kubenswrapper[4799]: I1129 04:54:29.312501 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74500694-328d-483c-8549-2bc90c72c811-kube-api-access-kmqvl" (OuterVolumeSpecName: "kube-api-access-kmqvl") pod "74500694-328d-483c-8549-2bc90c72c811" (UID: "74500694-328d-483c-8549-2bc90c72c811"). InnerVolumeSpecName "kube-api-access-kmqvl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:54:29 crc kubenswrapper[4799]: I1129 04:54:29.411997 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kmqvl\" (UniqueName: \"kubernetes.io/projected/74500694-328d-483c-8549-2bc90c72c811-kube-api-access-kmqvl\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:29 crc kubenswrapper[4799]: I1129 04:54:29.412051 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/74500694-328d-483c-8549-2bc90c72c811-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:29 crc kubenswrapper[4799]: I1129 04:54:29.859078 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1788e4d3-8c2a-4d06-b746-c0fb12514148","Type":"ContainerStarted","Data":"7abd82f4581f37354e1f18a9c0d46bf467ffbb2b06cbe83edec2915ef5bd8a74"} Nov 29 04:54:29 crc kubenswrapper[4799]: I1129 04:54:29.859845 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 29 04:54:29 crc kubenswrapper[4799]: I1129 04:54:29.865375 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-66ef-account-create-update-mcd96" Nov 29 04:54:29 crc kubenswrapper[4799]: I1129 04:54:29.865374 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-66ef-account-create-update-mcd96" event={"ID":"74500694-328d-483c-8549-2bc90c72c811","Type":"ContainerDied","Data":"5546b37350112c0c7c5948eae46937acf3e343ca5f0ea9f84e51111a934a2612"} Nov 29 04:54:29 crc kubenswrapper[4799]: I1129 04:54:29.865484 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5546b37350112c0c7c5948eae46937acf3e343ca5f0ea9f84e51111a934a2612" Nov 29 04:54:29 crc kubenswrapper[4799]: I1129 04:54:29.867734 4799 generic.go:334] "Generic (PLEG): container finished" podID="81726039-72e9-455a-b4fb-e16184acdc47" containerID="a72d02dc2cf8aa7523907bde01f61568e9db9f50c267ea96039aa2db0a285b08" exitCode=0 Nov 29 04:54:29 crc kubenswrapper[4799]: I1129 04:54:29.867757 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"81726039-72e9-455a-b4fb-e16184acdc47","Type":"ContainerDied","Data":"a72d02dc2cf8aa7523907bde01f61568e9db9f50c267ea96039aa2db0a285b08"} Nov 29 04:54:29 crc kubenswrapper[4799]: I1129 04:54:29.912397 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.361959671 podStartE2EDuration="1m19.912360366s" podCreationTimestamp="2025-11-29 04:53:10 +0000 UTC" firstStartedPulling="2025-11-29 04:53:12.384473665 +0000 UTC m=+868.027404055" lastFinishedPulling="2025-11-29 04:53:54.93487435 +0000 UTC m=+910.577804750" observedRunningTime="2025-11-29 04:54:29.89811072 +0000 UTC m=+945.541041160" watchObservedRunningTime="2025-11-29 04:54:29.912360366 +0000 UTC m=+945.555290776" Nov 29 04:54:30 crc kubenswrapper[4799]: I1129 04:54:30.672570 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89b0a912-606b-4d5d-bd40-09f89f2db649" path="/var/lib/kubelet/pods/89b0a912-606b-4d5d-bd40-09f89f2db649/volumes" Nov 29 04:54:30 crc kubenswrapper[4799]: I1129 04:54:30.881293 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"81726039-72e9-455a-b4fb-e16184acdc47","Type":"ContainerStarted","Data":"037ee3bbc021dfbf6239503e300c61dac479169025f6c1c3911be93913381a66"} Nov 29 04:54:30 crc kubenswrapper[4799]: I1129 04:54:30.882098 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:54:30 crc kubenswrapper[4799]: I1129 04:54:30.917187 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=38.330934428 podStartE2EDuration="1m20.916960764s" podCreationTimestamp="2025-11-29 04:53:10 +0000 UTC" firstStartedPulling="2025-11-29 04:53:12.698995164 +0000 UTC m=+868.341925564" lastFinishedPulling="2025-11-29 04:53:55.28502149 +0000 UTC m=+910.927951900" observedRunningTime="2025-11-29 04:54:30.913868521 +0000 UTC m=+946.556798941" watchObservedRunningTime="2025-11-29 04:54:30.916960764 +0000 UTC m=+946.559891164" Nov 29 04:54:31 crc kubenswrapper[4799]: I1129 04:54:31.527453 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-kkdw8" podUID="1d3d81e5-591a-469d-9851-86271162d455" containerName="ovn-controller" probeResult="failure" output=< Nov 29 04:54:31 crc kubenswrapper[4799]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 29 04:54:31 crc kubenswrapper[4799]: > Nov 29 04:54:31 crc kubenswrapper[4799]: I1129 04:54:31.637844 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-bkkst" Nov 29 04:54:31 crc kubenswrapper[4799]: I1129 04:54:31.638937 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-bkkst" Nov 29 04:54:31 crc kubenswrapper[4799]: I1129 04:54:31.850334 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-kkdw8-config-zqvjc"] Nov 29 04:54:31 crc kubenswrapper[4799]: E1129 04:54:31.850872 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e8bbbf6-43e4-4cd2-a3cd-326d3d11009f" containerName="mariadb-database-create" Nov 29 04:54:31 crc kubenswrapper[4799]: I1129 04:54:31.850890 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e8bbbf6-43e4-4cd2-a3cd-326d3d11009f" containerName="mariadb-database-create" Nov 29 04:54:31 crc kubenswrapper[4799]: E1129 04:54:31.850905 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89b0a912-606b-4d5d-bd40-09f89f2db649" containerName="extract-utilities" Nov 29 04:54:31 crc kubenswrapper[4799]: I1129 04:54:31.850914 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="89b0a912-606b-4d5d-bd40-09f89f2db649" containerName="extract-utilities" Nov 29 04:54:31 crc kubenswrapper[4799]: E1129 04:54:31.850926 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89b0a912-606b-4d5d-bd40-09f89f2db649" containerName="registry-server" Nov 29 04:54:31 crc kubenswrapper[4799]: I1129 04:54:31.850934 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="89b0a912-606b-4d5d-bd40-09f89f2db649" containerName="registry-server" Nov 29 04:54:31 crc kubenswrapper[4799]: E1129 04:54:31.850944 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f67bb6d3-7729-467b-9bab-c3aba0a555ff" containerName="mariadb-database-create" Nov 29 04:54:31 crc kubenswrapper[4799]: I1129 04:54:31.850951 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f67bb6d3-7729-467b-9bab-c3aba0a555ff" containerName="mariadb-database-create" Nov 29 04:54:31 crc kubenswrapper[4799]: E1129 04:54:31.850959 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74500694-328d-483c-8549-2bc90c72c811" containerName="mariadb-account-create-update" Nov 29 04:54:31 crc kubenswrapper[4799]: I1129 04:54:31.850969 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="74500694-328d-483c-8549-2bc90c72c811" containerName="mariadb-account-create-update" Nov 29 04:54:31 crc kubenswrapper[4799]: E1129 04:54:31.850992 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89b0a912-606b-4d5d-bd40-09f89f2db649" containerName="extract-content" Nov 29 04:54:31 crc kubenswrapper[4799]: I1129 04:54:31.850998 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="89b0a912-606b-4d5d-bd40-09f89f2db649" containerName="extract-content" Nov 29 04:54:31 crc kubenswrapper[4799]: E1129 04:54:31.851010 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58450435-f88c-4ad2-abd9-ff0c1c183527" containerName="mariadb-account-create-update" Nov 29 04:54:31 crc kubenswrapper[4799]: I1129 04:54:31.851018 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="58450435-f88c-4ad2-abd9-ff0c1c183527" containerName="mariadb-account-create-update" Nov 29 04:54:31 crc kubenswrapper[4799]: E1129 04:54:31.851033 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b3734a5-609e-4ffc-ac55-1d9fb1e699cc" containerName="dnsmasq-dns" Nov 29 04:54:31 crc kubenswrapper[4799]: I1129 04:54:31.851040 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b3734a5-609e-4ffc-ac55-1d9fb1e699cc" containerName="dnsmasq-dns" Nov 29 04:54:31 crc kubenswrapper[4799]: E1129 04:54:31.851062 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b3734a5-609e-4ffc-ac55-1d9fb1e699cc" containerName="init" Nov 29 04:54:31 crc kubenswrapper[4799]: I1129 04:54:31.851069 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b3734a5-609e-4ffc-ac55-1d9fb1e699cc" containerName="init" Nov 29 04:54:31 crc kubenswrapper[4799]: I1129 04:54:31.851274 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="f67bb6d3-7729-467b-9bab-c3aba0a555ff" containerName="mariadb-database-create" Nov 29 04:54:31 crc kubenswrapper[4799]: I1129 04:54:31.851290 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="89b0a912-606b-4d5d-bd40-09f89f2db649" containerName="registry-server" Nov 29 04:54:31 crc kubenswrapper[4799]: I1129 04:54:31.851299 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e8bbbf6-43e4-4cd2-a3cd-326d3d11009f" containerName="mariadb-database-create" Nov 29 04:54:31 crc kubenswrapper[4799]: I1129 04:54:31.851309 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="74500694-328d-483c-8549-2bc90c72c811" containerName="mariadb-account-create-update" Nov 29 04:54:31 crc kubenswrapper[4799]: I1129 04:54:31.851321 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="58450435-f88c-4ad2-abd9-ff0c1c183527" containerName="mariadb-account-create-update" Nov 29 04:54:31 crc kubenswrapper[4799]: I1129 04:54:31.851333 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b3734a5-609e-4ffc-ac55-1d9fb1e699cc" containerName="dnsmasq-dns" Nov 29 04:54:31 crc kubenswrapper[4799]: I1129 04:54:31.852114 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-kkdw8-config-zqvjc" Nov 29 04:54:31 crc kubenswrapper[4799]: I1129 04:54:31.855316 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 29 04:54:31 crc kubenswrapper[4799]: I1129 04:54:31.861575 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-kkdw8-config-zqvjc"] Nov 29 04:54:31 crc kubenswrapper[4799]: I1129 04:54:31.971877 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/631d5b8e-6703-4cee-96d7-4d014e38112e-var-run\") pod \"ovn-controller-kkdw8-config-zqvjc\" (UID: \"631d5b8e-6703-4cee-96d7-4d014e38112e\") " pod="openstack/ovn-controller-kkdw8-config-zqvjc" Nov 29 04:54:31 crc kubenswrapper[4799]: I1129 04:54:31.971960 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdrc8\" (UniqueName: \"kubernetes.io/projected/631d5b8e-6703-4cee-96d7-4d014e38112e-kube-api-access-xdrc8\") pod \"ovn-controller-kkdw8-config-zqvjc\" (UID: \"631d5b8e-6703-4cee-96d7-4d014e38112e\") " pod="openstack/ovn-controller-kkdw8-config-zqvjc" Nov 29 04:54:31 crc kubenswrapper[4799]: I1129 04:54:31.971988 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/631d5b8e-6703-4cee-96d7-4d014e38112e-scripts\") pod \"ovn-controller-kkdw8-config-zqvjc\" (UID: \"631d5b8e-6703-4cee-96d7-4d014e38112e\") " pod="openstack/ovn-controller-kkdw8-config-zqvjc" Nov 29 04:54:31 crc kubenswrapper[4799]: I1129 04:54:31.972173 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/631d5b8e-6703-4cee-96d7-4d014e38112e-additional-scripts\") pod \"ovn-controller-kkdw8-config-zqvjc\" (UID: \"631d5b8e-6703-4cee-96d7-4d014e38112e\") " pod="openstack/ovn-controller-kkdw8-config-zqvjc" Nov 29 04:54:31 crc kubenswrapper[4799]: I1129 04:54:31.972405 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/631d5b8e-6703-4cee-96d7-4d014e38112e-var-log-ovn\") pod \"ovn-controller-kkdw8-config-zqvjc\" (UID: \"631d5b8e-6703-4cee-96d7-4d014e38112e\") " pod="openstack/ovn-controller-kkdw8-config-zqvjc" Nov 29 04:54:31 crc kubenswrapper[4799]: I1129 04:54:31.972458 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/631d5b8e-6703-4cee-96d7-4d014e38112e-var-run-ovn\") pod \"ovn-controller-kkdw8-config-zqvjc\" (UID: \"631d5b8e-6703-4cee-96d7-4d014e38112e\") " pod="openstack/ovn-controller-kkdw8-config-zqvjc" Nov 29 04:54:32 crc kubenswrapper[4799]: I1129 04:54:32.075113 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/631d5b8e-6703-4cee-96d7-4d014e38112e-var-run-ovn\") pod \"ovn-controller-kkdw8-config-zqvjc\" (UID: \"631d5b8e-6703-4cee-96d7-4d014e38112e\") " pod="openstack/ovn-controller-kkdw8-config-zqvjc" Nov 29 04:54:32 crc kubenswrapper[4799]: I1129 04:54:32.075257 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/631d5b8e-6703-4cee-96d7-4d014e38112e-var-run\") pod \"ovn-controller-kkdw8-config-zqvjc\" (UID: \"631d5b8e-6703-4cee-96d7-4d014e38112e\") " pod="openstack/ovn-controller-kkdw8-config-zqvjc" Nov 29 04:54:32 crc kubenswrapper[4799]: I1129 04:54:32.075309 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdrc8\" (UniqueName: \"kubernetes.io/projected/631d5b8e-6703-4cee-96d7-4d014e38112e-kube-api-access-xdrc8\") pod \"ovn-controller-kkdw8-config-zqvjc\" (UID: \"631d5b8e-6703-4cee-96d7-4d014e38112e\") " pod="openstack/ovn-controller-kkdw8-config-zqvjc" Nov 29 04:54:32 crc kubenswrapper[4799]: I1129 04:54:32.075331 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/631d5b8e-6703-4cee-96d7-4d014e38112e-scripts\") pod \"ovn-controller-kkdw8-config-zqvjc\" (UID: \"631d5b8e-6703-4cee-96d7-4d014e38112e\") " pod="openstack/ovn-controller-kkdw8-config-zqvjc" Nov 29 04:54:32 crc kubenswrapper[4799]: I1129 04:54:32.075366 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/631d5b8e-6703-4cee-96d7-4d014e38112e-additional-scripts\") pod \"ovn-controller-kkdw8-config-zqvjc\" (UID: \"631d5b8e-6703-4cee-96d7-4d014e38112e\") " pod="openstack/ovn-controller-kkdw8-config-zqvjc" Nov 29 04:54:32 crc kubenswrapper[4799]: I1129 04:54:32.075503 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/631d5b8e-6703-4cee-96d7-4d014e38112e-var-run-ovn\") pod \"ovn-controller-kkdw8-config-zqvjc\" (UID: \"631d5b8e-6703-4cee-96d7-4d014e38112e\") " pod="openstack/ovn-controller-kkdw8-config-zqvjc" Nov 29 04:54:32 crc kubenswrapper[4799]: I1129 04:54:32.075584 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/631d5b8e-6703-4cee-96d7-4d014e38112e-var-log-ovn\") pod \"ovn-controller-kkdw8-config-zqvjc\" (UID: \"631d5b8e-6703-4cee-96d7-4d014e38112e\") " pod="openstack/ovn-controller-kkdw8-config-zqvjc" Nov 29 04:54:32 crc kubenswrapper[4799]: I1129 04:54:32.075661 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/631d5b8e-6703-4cee-96d7-4d014e38112e-var-log-ovn\") pod \"ovn-controller-kkdw8-config-zqvjc\" (UID: \"631d5b8e-6703-4cee-96d7-4d014e38112e\") " pod="openstack/ovn-controller-kkdw8-config-zqvjc" Nov 29 04:54:32 crc kubenswrapper[4799]: I1129 04:54:32.075823 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/631d5b8e-6703-4cee-96d7-4d014e38112e-var-run\") pod \"ovn-controller-kkdw8-config-zqvjc\" (UID: \"631d5b8e-6703-4cee-96d7-4d014e38112e\") " pod="openstack/ovn-controller-kkdw8-config-zqvjc" Nov 29 04:54:32 crc kubenswrapper[4799]: I1129 04:54:32.076848 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/631d5b8e-6703-4cee-96d7-4d014e38112e-additional-scripts\") pod \"ovn-controller-kkdw8-config-zqvjc\" (UID: \"631d5b8e-6703-4cee-96d7-4d014e38112e\") " pod="openstack/ovn-controller-kkdw8-config-zqvjc" Nov 29 04:54:32 crc kubenswrapper[4799]: I1129 04:54:32.078258 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/631d5b8e-6703-4cee-96d7-4d014e38112e-scripts\") pod \"ovn-controller-kkdw8-config-zqvjc\" (UID: \"631d5b8e-6703-4cee-96d7-4d014e38112e\") " pod="openstack/ovn-controller-kkdw8-config-zqvjc" Nov 29 04:54:32 crc kubenswrapper[4799]: I1129 04:54:32.095648 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdrc8\" (UniqueName: \"kubernetes.io/projected/631d5b8e-6703-4cee-96d7-4d014e38112e-kube-api-access-xdrc8\") pod \"ovn-controller-kkdw8-config-zqvjc\" (UID: \"631d5b8e-6703-4cee-96d7-4d014e38112e\") " pod="openstack/ovn-controller-kkdw8-config-zqvjc" Nov 29 04:54:32 crc kubenswrapper[4799]: I1129 04:54:32.173575 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-kkdw8-config-zqvjc" Nov 29 04:54:32 crc kubenswrapper[4799]: I1129 04:54:32.854868 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-kkdw8-config-zqvjc"] Nov 29 04:54:32 crc kubenswrapper[4799]: W1129 04:54:32.860859 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod631d5b8e_6703_4cee_96d7_4d014e38112e.slice/crio-850eda75329fd448cb27b730a4674ccdc44b020b42ddb2f067a3e6558f8a8e62 WatchSource:0}: Error finding container 850eda75329fd448cb27b730a4674ccdc44b020b42ddb2f067a3e6558f8a8e62: Status 404 returned error can't find the container with id 850eda75329fd448cb27b730a4674ccdc44b020b42ddb2f067a3e6558f8a8e62 Nov 29 04:54:32 crc kubenswrapper[4799]: I1129 04:54:32.916660 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-kkdw8-config-zqvjc" event={"ID":"631d5b8e-6703-4cee-96d7-4d014e38112e","Type":"ContainerStarted","Data":"850eda75329fd448cb27b730a4674ccdc44b020b42ddb2f067a3e6558f8a8e62"} Nov 29 04:54:33 crc kubenswrapper[4799]: I1129 04:54:33.276799 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-24kgw" Nov 29 04:54:33 crc kubenswrapper[4799]: I1129 04:54:33.330372 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-24kgw"] Nov 29 04:54:33 crc kubenswrapper[4799]: I1129 04:54:33.956994 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-kkdw8-config-zqvjc" event={"ID":"631d5b8e-6703-4cee-96d7-4d014e38112e","Type":"ContainerStarted","Data":"d31ed74a4ffe842e5ce4e12afae41c8602e5381cb90d93d341a7b3e77a97a520"} Nov 29 04:54:33 crc kubenswrapper[4799]: I1129 04:54:33.957586 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-24kgw" podUID="36215eda-2af7-4388-992f-b31807c9d423" containerName="registry-server" containerID="cri-o://fb79e68053b18642c9521ab4a168eee0942b1899c610dfa8021255d33f5d6189" gracePeriod=2 Nov 29 04:54:33 crc kubenswrapper[4799]: I1129 04:54:33.975705 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-kkdw8-config-zqvjc" podStartSLOduration=2.97567972 podStartE2EDuration="2.97567972s" podCreationTimestamp="2025-11-29 04:54:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:54:33.974983193 +0000 UTC m=+949.617913603" watchObservedRunningTime="2025-11-29 04:54:33.97567972 +0000 UTC m=+949.618610120" Nov 29 04:54:34 crc kubenswrapper[4799]: I1129 04:54:34.990672 4799 generic.go:334] "Generic (PLEG): container finished" podID="631d5b8e-6703-4cee-96d7-4d014e38112e" containerID="d31ed74a4ffe842e5ce4e12afae41c8602e5381cb90d93d341a7b3e77a97a520" exitCode=0 Nov 29 04:54:34 crc kubenswrapper[4799]: I1129 04:54:34.990859 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-kkdw8-config-zqvjc" event={"ID":"631d5b8e-6703-4cee-96d7-4d014e38112e","Type":"ContainerDied","Data":"d31ed74a4ffe842e5ce4e12afae41c8602e5381cb90d93d341a7b3e77a97a520"} Nov 29 04:54:34 crc kubenswrapper[4799]: I1129 04:54:34.996777 4799 generic.go:334] "Generic (PLEG): container finished" podID="36215eda-2af7-4388-992f-b31807c9d423" containerID="fb79e68053b18642c9521ab4a168eee0942b1899c610dfa8021255d33f5d6189" exitCode=0 Nov 29 04:54:34 crc kubenswrapper[4799]: I1129 04:54:34.996847 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-24kgw" event={"ID":"36215eda-2af7-4388-992f-b31807c9d423","Type":"ContainerDied","Data":"fb79e68053b18642c9521ab4a168eee0942b1899c610dfa8021255d33f5d6189"} Nov 29 04:54:35 crc kubenswrapper[4799]: I1129 04:54:35.420993 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Nov 29 04:54:36 crc kubenswrapper[4799]: I1129 04:54:36.527495 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-kkdw8" Nov 29 04:54:37 crc kubenswrapper[4799]: I1129 04:54:37.498432 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 04:54:37 crc kubenswrapper[4799]: I1129 04:54:37.498858 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 04:54:41 crc kubenswrapper[4799]: I1129 04:54:41.770134 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.052709 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-kkdw8-config-zqvjc" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.073840 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-kkdw8-config-zqvjc" event={"ID":"631d5b8e-6703-4cee-96d7-4d014e38112e","Type":"ContainerDied","Data":"850eda75329fd448cb27b730a4674ccdc44b020b42ddb2f067a3e6558f8a8e62"} Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.073901 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="850eda75329fd448cb27b730a4674ccdc44b020b42ddb2f067a3e6558f8a8e62" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.073982 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-kkdw8-config-zqvjc" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.089554 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xdrc8\" (UniqueName: \"kubernetes.io/projected/631d5b8e-6703-4cee-96d7-4d014e38112e-kube-api-access-xdrc8\") pod \"631d5b8e-6703-4cee-96d7-4d014e38112e\" (UID: \"631d5b8e-6703-4cee-96d7-4d014e38112e\") " Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.090016 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/631d5b8e-6703-4cee-96d7-4d014e38112e-scripts\") pod \"631d5b8e-6703-4cee-96d7-4d014e38112e\" (UID: \"631d5b8e-6703-4cee-96d7-4d014e38112e\") " Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.090055 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/631d5b8e-6703-4cee-96d7-4d014e38112e-var-log-ovn\") pod \"631d5b8e-6703-4cee-96d7-4d014e38112e\" (UID: \"631d5b8e-6703-4cee-96d7-4d014e38112e\") " Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.090081 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/631d5b8e-6703-4cee-96d7-4d014e38112e-var-run-ovn\") pod \"631d5b8e-6703-4cee-96d7-4d014e38112e\" (UID: \"631d5b8e-6703-4cee-96d7-4d014e38112e\") " Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.090116 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/631d5b8e-6703-4cee-96d7-4d014e38112e-additional-scripts\") pod \"631d5b8e-6703-4cee-96d7-4d014e38112e\" (UID: \"631d5b8e-6703-4cee-96d7-4d014e38112e\") " Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.090140 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/631d5b8e-6703-4cee-96d7-4d014e38112e-var-run\") pod \"631d5b8e-6703-4cee-96d7-4d014e38112e\" (UID: \"631d5b8e-6703-4cee-96d7-4d014e38112e\") " Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.090442 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/631d5b8e-6703-4cee-96d7-4d014e38112e-var-run" (OuterVolumeSpecName: "var-run") pod "631d5b8e-6703-4cee-96d7-4d014e38112e" (UID: "631d5b8e-6703-4cee-96d7-4d014e38112e"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.090430 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/631d5b8e-6703-4cee-96d7-4d014e38112e-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "631d5b8e-6703-4cee-96d7-4d014e38112e" (UID: "631d5b8e-6703-4cee-96d7-4d014e38112e"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.090477 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/631d5b8e-6703-4cee-96d7-4d014e38112e-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "631d5b8e-6703-4cee-96d7-4d014e38112e" (UID: "631d5b8e-6703-4cee-96d7-4d014e38112e"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.091210 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/631d5b8e-6703-4cee-96d7-4d014e38112e-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "631d5b8e-6703-4cee-96d7-4d014e38112e" (UID: "631d5b8e-6703-4cee-96d7-4d014e38112e"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.091483 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/631d5b8e-6703-4cee-96d7-4d014e38112e-scripts" (OuterVolumeSpecName: "scripts") pod "631d5b8e-6703-4cee-96d7-4d014e38112e" (UID: "631d5b8e-6703-4cee-96d7-4d014e38112e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.113170 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/631d5b8e-6703-4cee-96d7-4d014e38112e-kube-api-access-xdrc8" (OuterVolumeSpecName: "kube-api-access-xdrc8") pod "631d5b8e-6703-4cee-96d7-4d014e38112e" (UID: "631d5b8e-6703-4cee-96d7-4d014e38112e"). InnerVolumeSpecName "kube-api-access-xdrc8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.113989 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.180411 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-9vlv8"] Nov 29 04:54:42 crc kubenswrapper[4799]: E1129 04:54:42.185570 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="631d5b8e-6703-4cee-96d7-4d014e38112e" containerName="ovn-config" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.185600 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="631d5b8e-6703-4cee-96d7-4d014e38112e" containerName="ovn-config" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.185907 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="631d5b8e-6703-4cee-96d7-4d014e38112e" containerName="ovn-config" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.186702 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-9vlv8" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.193355 4799 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/631d5b8e-6703-4cee-96d7-4d014e38112e-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.193397 4799 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/631d5b8e-6703-4cee-96d7-4d014e38112e-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.193411 4799 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/631d5b8e-6703-4cee-96d7-4d014e38112e-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.193422 4799 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/631d5b8e-6703-4cee-96d7-4d014e38112e-var-run\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.193434 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xdrc8\" (UniqueName: \"kubernetes.io/projected/631d5b8e-6703-4cee-96d7-4d014e38112e-kube-api-access-xdrc8\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.193445 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/631d5b8e-6703-4cee-96d7-4d014e38112e-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.213935 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-9vlv8"] Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.295085 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1a762825-986d-4691-88bb-da6580cf83b1-operator-scripts\") pod \"barbican-db-create-9vlv8\" (UID: \"1a762825-986d-4691-88bb-da6580cf83b1\") " pod="openstack/barbican-db-create-9vlv8" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.295203 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kqq9\" (UniqueName: \"kubernetes.io/projected/1a762825-986d-4691-88bb-da6580cf83b1-kube-api-access-8kqq9\") pod \"barbican-db-create-9vlv8\" (UID: \"1a762825-986d-4691-88bb-da6580cf83b1\") " pod="openstack/barbican-db-create-9vlv8" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.365990 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-7da7-account-create-update-wr6xq"] Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.368004 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-7da7-account-create-update-wr6xq" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.370796 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.372414 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-7da7-account-create-update-wr6xq"] Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.396600 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kqq9\" (UniqueName: \"kubernetes.io/projected/1a762825-986d-4691-88bb-da6580cf83b1-kube-api-access-8kqq9\") pod \"barbican-db-create-9vlv8\" (UID: \"1a762825-986d-4691-88bb-da6580cf83b1\") " pod="openstack/barbican-db-create-9vlv8" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.396725 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1a762825-986d-4691-88bb-da6580cf83b1-operator-scripts\") pod \"barbican-db-create-9vlv8\" (UID: \"1a762825-986d-4691-88bb-da6580cf83b1\") " pod="openstack/barbican-db-create-9vlv8" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.397523 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1a762825-986d-4691-88bb-da6580cf83b1-operator-scripts\") pod \"barbican-db-create-9vlv8\" (UID: \"1a762825-986d-4691-88bb-da6580cf83b1\") " pod="openstack/barbican-db-create-9vlv8" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.443433 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-tv4bk"] Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.444646 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-tv4bk" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.447147 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kqq9\" (UniqueName: \"kubernetes.io/projected/1a762825-986d-4691-88bb-da6580cf83b1-kube-api-access-8kqq9\") pod \"barbican-db-create-9vlv8\" (UID: \"1a762825-986d-4691-88bb-da6580cf83b1\") " pod="openstack/barbican-db-create-9vlv8" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.481295 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-tv4bk"] Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.499225 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfjg8\" (UniqueName: \"kubernetes.io/projected/b5536f91-0fec-4581-8a6d-18b365e7978b-kube-api-access-mfjg8\") pod \"barbican-7da7-account-create-update-wr6xq\" (UID: \"b5536f91-0fec-4581-8a6d-18b365e7978b\") " pod="openstack/barbican-7da7-account-create-update-wr6xq" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.499377 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5536f91-0fec-4581-8a6d-18b365e7978b-operator-scripts\") pod \"barbican-7da7-account-create-update-wr6xq\" (UID: \"b5536f91-0fec-4581-8a6d-18b365e7978b\") " pod="openstack/barbican-7da7-account-create-update-wr6xq" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.548329 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-2f96-account-create-update-4gdd6"] Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.549661 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-2f96-account-create-update-4gdd6" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.550987 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-24kgw" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.553524 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.573121 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-2f96-account-create-update-4gdd6"] Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.585775 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-9vlv8" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.601782 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc7d764a-f94f-4f3a-bf98-643a323f7cc4-operator-scripts\") pod \"cinder-db-create-tv4bk\" (UID: \"fc7d764a-f94f-4f3a-bf98-643a323f7cc4\") " pod="openstack/cinder-db-create-tv4bk" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.601878 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfjg8\" (UniqueName: \"kubernetes.io/projected/b5536f91-0fec-4581-8a6d-18b365e7978b-kube-api-access-mfjg8\") pod \"barbican-7da7-account-create-update-wr6xq\" (UID: \"b5536f91-0fec-4581-8a6d-18b365e7978b\") " pod="openstack/barbican-7da7-account-create-update-wr6xq" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.601984 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mq5xs\" (UniqueName: \"kubernetes.io/projected/fc7d764a-f94f-4f3a-bf98-643a323f7cc4-kube-api-access-mq5xs\") pod \"cinder-db-create-tv4bk\" (UID: \"fc7d764a-f94f-4f3a-bf98-643a323f7cc4\") " pod="openstack/cinder-db-create-tv4bk" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.602022 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5536f91-0fec-4581-8a6d-18b365e7978b-operator-scripts\") pod \"barbican-7da7-account-create-update-wr6xq\" (UID: \"b5536f91-0fec-4581-8a6d-18b365e7978b\") " pod="openstack/barbican-7da7-account-create-update-wr6xq" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.602966 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5536f91-0fec-4581-8a6d-18b365e7978b-operator-scripts\") pod \"barbican-7da7-account-create-update-wr6xq\" (UID: \"b5536f91-0fec-4581-8a6d-18b365e7978b\") " pod="openstack/barbican-7da7-account-create-update-wr6xq" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.634740 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfjg8\" (UniqueName: \"kubernetes.io/projected/b5536f91-0fec-4581-8a6d-18b365e7978b-kube-api-access-mfjg8\") pod \"barbican-7da7-account-create-update-wr6xq\" (UID: \"b5536f91-0fec-4581-8a6d-18b365e7978b\") " pod="openstack/barbican-7da7-account-create-update-wr6xq" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.648060 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-665zk"] Nov 29 04:54:42 crc kubenswrapper[4799]: E1129 04:54:42.648528 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36215eda-2af7-4388-992f-b31807c9d423" containerName="registry-server" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.648543 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="36215eda-2af7-4388-992f-b31807c9d423" containerName="registry-server" Nov 29 04:54:42 crc kubenswrapper[4799]: E1129 04:54:42.648557 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36215eda-2af7-4388-992f-b31807c9d423" containerName="extract-content" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.648564 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="36215eda-2af7-4388-992f-b31807c9d423" containerName="extract-content" Nov 29 04:54:42 crc kubenswrapper[4799]: E1129 04:54:42.648584 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36215eda-2af7-4388-992f-b31807c9d423" containerName="extract-utilities" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.648591 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="36215eda-2af7-4388-992f-b31807c9d423" containerName="extract-utilities" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.648755 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="36215eda-2af7-4388-992f-b31807c9d423" containerName="registry-server" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.649588 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-665zk" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.690633 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-665zk"] Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.701489 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-7da7-account-create-update-wr6xq" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.704337 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36215eda-2af7-4388-992f-b31807c9d423-utilities\") pod \"36215eda-2af7-4388-992f-b31807c9d423\" (UID: \"36215eda-2af7-4388-992f-b31807c9d423\") " Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.704398 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36215eda-2af7-4388-992f-b31807c9d423-catalog-content\") pod \"36215eda-2af7-4388-992f-b31807c9d423\" (UID: \"36215eda-2af7-4388-992f-b31807c9d423\") " Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.704508 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7cv7\" (UniqueName: \"kubernetes.io/projected/36215eda-2af7-4388-992f-b31807c9d423-kube-api-access-h7cv7\") pod \"36215eda-2af7-4388-992f-b31807c9d423\" (UID: \"36215eda-2af7-4388-992f-b31807c9d423\") " Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.704783 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89p6t\" (UniqueName: \"kubernetes.io/projected/29be7ae1-328d-4e71-9796-f054b8f4686f-kube-api-access-89p6t\") pod \"cinder-2f96-account-create-update-4gdd6\" (UID: \"29be7ae1-328d-4e71-9796-f054b8f4686f\") " pod="openstack/cinder-2f96-account-create-update-4gdd6" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.704875 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mq5xs\" (UniqueName: \"kubernetes.io/projected/fc7d764a-f94f-4f3a-bf98-643a323f7cc4-kube-api-access-mq5xs\") pod \"cinder-db-create-tv4bk\" (UID: \"fc7d764a-f94f-4f3a-bf98-643a323f7cc4\") " pod="openstack/cinder-db-create-tv4bk" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.704933 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc7d764a-f94f-4f3a-bf98-643a323f7cc4-operator-scripts\") pod \"cinder-db-create-tv4bk\" (UID: \"fc7d764a-f94f-4f3a-bf98-643a323f7cc4\") " pod="openstack/cinder-db-create-tv4bk" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.704955 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/29be7ae1-328d-4e71-9796-f054b8f4686f-operator-scripts\") pod \"cinder-2f96-account-create-update-4gdd6\" (UID: \"29be7ae1-328d-4e71-9796-f054b8f4686f\") " pod="openstack/cinder-2f96-account-create-update-4gdd6" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.706839 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc7d764a-f94f-4f3a-bf98-643a323f7cc4-operator-scripts\") pod \"cinder-db-create-tv4bk\" (UID: \"fc7d764a-f94f-4f3a-bf98-643a323f7cc4\") " pod="openstack/cinder-db-create-tv4bk" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.707590 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36215eda-2af7-4388-992f-b31807c9d423-utilities" (OuterVolumeSpecName: "utilities") pod "36215eda-2af7-4388-992f-b31807c9d423" (UID: "36215eda-2af7-4388-992f-b31807c9d423"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.715135 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36215eda-2af7-4388-992f-b31807c9d423-kube-api-access-h7cv7" (OuterVolumeSpecName: "kube-api-access-h7cv7") pod "36215eda-2af7-4388-992f-b31807c9d423" (UID: "36215eda-2af7-4388-992f-b31807c9d423"). InnerVolumeSpecName "kube-api-access-h7cv7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.723590 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-gjbqp"] Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.726903 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-gjbqp" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.734751 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.735186 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.735429 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.735573 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-85vg6" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.749013 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36215eda-2af7-4388-992f-b31807c9d423-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "36215eda-2af7-4388-992f-b31807c9d423" (UID: "36215eda-2af7-4388-992f-b31807c9d423"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.749174 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mq5xs\" (UniqueName: \"kubernetes.io/projected/fc7d764a-f94f-4f3a-bf98-643a323f7cc4-kube-api-access-mq5xs\") pod \"cinder-db-create-tv4bk\" (UID: \"fc7d764a-f94f-4f3a-bf98-643a323f7cc4\") " pod="openstack/cinder-db-create-tv4bk" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.758900 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-gjbqp"] Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.788370 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-tv4bk" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.797877 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7451-account-create-update-r6m7w"] Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.799540 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7451-account-create-update-r6m7w" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.806216 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.807944 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5832499-91f5-48ca-9502-0931ae1ece9e-config-data\") pod \"keystone-db-sync-gjbqp\" (UID: \"b5832499-91f5-48ca-9502-0931ae1ece9e\") " pod="openstack/keystone-db-sync-gjbqp" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.808058 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/29be7ae1-328d-4e71-9796-f054b8f4686f-operator-scripts\") pod \"cinder-2f96-account-create-update-4gdd6\" (UID: \"29be7ae1-328d-4e71-9796-f054b8f4686f\") " pod="openstack/cinder-2f96-account-create-update-4gdd6" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.808110 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5832499-91f5-48ca-9502-0931ae1ece9e-combined-ca-bundle\") pod \"keystone-db-sync-gjbqp\" (UID: \"b5832499-91f5-48ca-9502-0931ae1ece9e\") " pod="openstack/keystone-db-sync-gjbqp" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.808131 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ac99f10b-e4cb-4847-a60b-69306ff5e8ca-operator-scripts\") pod \"neutron-db-create-665zk\" (UID: \"ac99f10b-e4cb-4847-a60b-69306ff5e8ca\") " pod="openstack/neutron-db-create-665zk" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.808156 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89p6t\" (UniqueName: \"kubernetes.io/projected/29be7ae1-328d-4e71-9796-f054b8f4686f-kube-api-access-89p6t\") pod \"cinder-2f96-account-create-update-4gdd6\" (UID: \"29be7ae1-328d-4e71-9796-f054b8f4686f\") " pod="openstack/cinder-2f96-account-create-update-4gdd6" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.808185 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppp45\" (UniqueName: \"kubernetes.io/projected/b5832499-91f5-48ca-9502-0931ae1ece9e-kube-api-access-ppp45\") pod \"keystone-db-sync-gjbqp\" (UID: \"b5832499-91f5-48ca-9502-0931ae1ece9e\") " pod="openstack/keystone-db-sync-gjbqp" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.808217 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbpr5\" (UniqueName: \"kubernetes.io/projected/ac99f10b-e4cb-4847-a60b-69306ff5e8ca-kube-api-access-fbpr5\") pod \"neutron-db-create-665zk\" (UID: \"ac99f10b-e4cb-4847-a60b-69306ff5e8ca\") " pod="openstack/neutron-db-create-665zk" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.808296 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36215eda-2af7-4388-992f-b31807c9d423-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.808313 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36215eda-2af7-4388-992f-b31807c9d423-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.808327 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7cv7\" (UniqueName: \"kubernetes.io/projected/36215eda-2af7-4388-992f-b31807c9d423-kube-api-access-h7cv7\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.809211 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/29be7ae1-328d-4e71-9796-f054b8f4686f-operator-scripts\") pod \"cinder-2f96-account-create-update-4gdd6\" (UID: \"29be7ae1-328d-4e71-9796-f054b8f4686f\") " pod="openstack/cinder-2f96-account-create-update-4gdd6" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.836376 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7451-account-create-update-r6m7w"] Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.855657 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89p6t\" (UniqueName: \"kubernetes.io/projected/29be7ae1-328d-4e71-9796-f054b8f4686f-kube-api-access-89p6t\") pod \"cinder-2f96-account-create-update-4gdd6\" (UID: \"29be7ae1-328d-4e71-9796-f054b8f4686f\") " pod="openstack/cinder-2f96-account-create-update-4gdd6" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.884123 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-2f96-account-create-update-4gdd6" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.910913 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbpr5\" (UniqueName: \"kubernetes.io/projected/ac99f10b-e4cb-4847-a60b-69306ff5e8ca-kube-api-access-fbpr5\") pod \"neutron-db-create-665zk\" (UID: \"ac99f10b-e4cb-4847-a60b-69306ff5e8ca\") " pod="openstack/neutron-db-create-665zk" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.911005 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrdtt\" (UniqueName: \"kubernetes.io/projected/ee16aa26-5b84-447d-8a02-7a4199451cc3-kube-api-access-zrdtt\") pod \"neutron-7451-account-create-update-r6m7w\" (UID: \"ee16aa26-5b84-447d-8a02-7a4199451cc3\") " pod="openstack/neutron-7451-account-create-update-r6m7w" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.911072 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5832499-91f5-48ca-9502-0931ae1ece9e-config-data\") pod \"keystone-db-sync-gjbqp\" (UID: \"b5832499-91f5-48ca-9502-0931ae1ece9e\") " pod="openstack/keystone-db-sync-gjbqp" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.911130 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5832499-91f5-48ca-9502-0931ae1ece9e-combined-ca-bundle\") pod \"keystone-db-sync-gjbqp\" (UID: \"b5832499-91f5-48ca-9502-0931ae1ece9e\") " pod="openstack/keystone-db-sync-gjbqp" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.911150 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ac99f10b-e4cb-4847-a60b-69306ff5e8ca-operator-scripts\") pod \"neutron-db-create-665zk\" (UID: \"ac99f10b-e4cb-4847-a60b-69306ff5e8ca\") " pod="openstack/neutron-db-create-665zk" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.911173 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee16aa26-5b84-447d-8a02-7a4199451cc3-operator-scripts\") pod \"neutron-7451-account-create-update-r6m7w\" (UID: \"ee16aa26-5b84-447d-8a02-7a4199451cc3\") " pod="openstack/neutron-7451-account-create-update-r6m7w" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.911195 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppp45\" (UniqueName: \"kubernetes.io/projected/b5832499-91f5-48ca-9502-0931ae1ece9e-kube-api-access-ppp45\") pod \"keystone-db-sync-gjbqp\" (UID: \"b5832499-91f5-48ca-9502-0931ae1ece9e\") " pod="openstack/keystone-db-sync-gjbqp" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.913741 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ac99f10b-e4cb-4847-a60b-69306ff5e8ca-operator-scripts\") pod \"neutron-db-create-665zk\" (UID: \"ac99f10b-e4cb-4847-a60b-69306ff5e8ca\") " pod="openstack/neutron-db-create-665zk" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.923750 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5832499-91f5-48ca-9502-0931ae1ece9e-config-data\") pod \"keystone-db-sync-gjbqp\" (UID: \"b5832499-91f5-48ca-9502-0931ae1ece9e\") " pod="openstack/keystone-db-sync-gjbqp" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.930530 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5832499-91f5-48ca-9502-0931ae1ece9e-combined-ca-bundle\") pod \"keystone-db-sync-gjbqp\" (UID: \"b5832499-91f5-48ca-9502-0931ae1ece9e\") " pod="openstack/keystone-db-sync-gjbqp" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.981674 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbpr5\" (UniqueName: \"kubernetes.io/projected/ac99f10b-e4cb-4847-a60b-69306ff5e8ca-kube-api-access-fbpr5\") pod \"neutron-db-create-665zk\" (UID: \"ac99f10b-e4cb-4847-a60b-69306ff5e8ca\") " pod="openstack/neutron-db-create-665zk" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.984071 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-665zk" Nov 29 04:54:42 crc kubenswrapper[4799]: I1129 04:54:42.985081 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppp45\" (UniqueName: \"kubernetes.io/projected/b5832499-91f5-48ca-9502-0931ae1ece9e-kube-api-access-ppp45\") pod \"keystone-db-sync-gjbqp\" (UID: \"b5832499-91f5-48ca-9502-0931ae1ece9e\") " pod="openstack/keystone-db-sync-gjbqp" Nov 29 04:54:43 crc kubenswrapper[4799]: I1129 04:54:43.021590 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee16aa26-5b84-447d-8a02-7a4199451cc3-operator-scripts\") pod \"neutron-7451-account-create-update-r6m7w\" (UID: \"ee16aa26-5b84-447d-8a02-7a4199451cc3\") " pod="openstack/neutron-7451-account-create-update-r6m7w" Nov 29 04:54:43 crc kubenswrapper[4799]: I1129 04:54:43.021688 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrdtt\" (UniqueName: \"kubernetes.io/projected/ee16aa26-5b84-447d-8a02-7a4199451cc3-kube-api-access-zrdtt\") pod \"neutron-7451-account-create-update-r6m7w\" (UID: \"ee16aa26-5b84-447d-8a02-7a4199451cc3\") " pod="openstack/neutron-7451-account-create-update-r6m7w" Nov 29 04:54:43 crc kubenswrapper[4799]: I1129 04:54:43.042499 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrdtt\" (UniqueName: \"kubernetes.io/projected/ee16aa26-5b84-447d-8a02-7a4199451cc3-kube-api-access-zrdtt\") pod \"neutron-7451-account-create-update-r6m7w\" (UID: \"ee16aa26-5b84-447d-8a02-7a4199451cc3\") " pod="openstack/neutron-7451-account-create-update-r6m7w" Nov 29 04:54:43 crc kubenswrapper[4799]: I1129 04:54:43.052374 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee16aa26-5b84-447d-8a02-7a4199451cc3-operator-scripts\") pod \"neutron-7451-account-create-update-r6m7w\" (UID: \"ee16aa26-5b84-447d-8a02-7a4199451cc3\") " pod="openstack/neutron-7451-account-create-update-r6m7w" Nov 29 04:54:43 crc kubenswrapper[4799]: I1129 04:54:43.058532 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-gjbqp" Nov 29 04:54:43 crc kubenswrapper[4799]: I1129 04:54:43.131159 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7451-account-create-update-r6m7w" Nov 29 04:54:43 crc kubenswrapper[4799]: I1129 04:54:43.201722 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-24kgw" event={"ID":"36215eda-2af7-4388-992f-b31807c9d423","Type":"ContainerDied","Data":"c20704c6b3ad06b103ea0f1de65eee4ad72a2ee7519799abbef097d50fe22743"} Nov 29 04:54:43 crc kubenswrapper[4799]: I1129 04:54:43.202129 4799 scope.go:117] "RemoveContainer" containerID="fb79e68053b18642c9521ab4a168eee0942b1899c610dfa8021255d33f5d6189" Nov 29 04:54:43 crc kubenswrapper[4799]: I1129 04:54:43.202363 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-24kgw" Nov 29 04:54:43 crc kubenswrapper[4799]: I1129 04:54:43.315823 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-9vlv8"] Nov 29 04:54:43 crc kubenswrapper[4799]: W1129 04:54:43.331948 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a762825_986d_4691_88bb_da6580cf83b1.slice/crio-f4902acb6f541b43569ee0d46a21153657f536232e4a91fcc3fb0f55f4791b96 WatchSource:0}: Error finding container f4902acb6f541b43569ee0d46a21153657f536232e4a91fcc3fb0f55f4791b96: Status 404 returned error can't find the container with id f4902acb6f541b43569ee0d46a21153657f536232e4a91fcc3fb0f55f4791b96 Nov 29 04:54:43 crc kubenswrapper[4799]: I1129 04:54:43.336234 4799 scope.go:117] "RemoveContainer" containerID="484a1d6e238a655ab1bf8dc288912082c32099706e4b71564d8a42fc9796e7b8" Nov 29 04:54:43 crc kubenswrapper[4799]: I1129 04:54:43.337303 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-kkdw8-config-zqvjc"] Nov 29 04:54:43 crc kubenswrapper[4799]: I1129 04:54:43.390692 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-kkdw8-config-zqvjc"] Nov 29 04:54:43 crc kubenswrapper[4799]: I1129 04:54:43.408113 4799 scope.go:117] "RemoveContainer" containerID="baba2e8023cbab4935c7039e780d17ac58c69198100aab1af217d7a18c88876f" Nov 29 04:54:43 crc kubenswrapper[4799]: I1129 04:54:43.431769 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-24kgw"] Nov 29 04:54:43 crc kubenswrapper[4799]: I1129 04:54:43.454553 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-24kgw"] Nov 29 04:54:43 crc kubenswrapper[4799]: I1129 04:54:43.639522 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-tv4bk"] Nov 29 04:54:43 crc kubenswrapper[4799]: W1129 04:54:43.651657 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfc7d764a_f94f_4f3a_bf98_643a323f7cc4.slice/crio-957d7bdec552d77a16d027232ac19c3bd61b08b6cf5ac45e001ef338191ea537 WatchSource:0}: Error finding container 957d7bdec552d77a16d027232ac19c3bd61b08b6cf5ac45e001ef338191ea537: Status 404 returned error can't find the container with id 957d7bdec552d77a16d027232ac19c3bd61b08b6cf5ac45e001ef338191ea537 Nov 29 04:54:43 crc kubenswrapper[4799]: I1129 04:54:43.912321 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-7da7-account-create-update-wr6xq"] Nov 29 04:54:43 crc kubenswrapper[4799]: I1129 04:54:43.934532 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-665zk"] Nov 29 04:54:43 crc kubenswrapper[4799]: W1129 04:54:43.940110 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podac99f10b_e4cb_4847_a60b_69306ff5e8ca.slice/crio-11bb68c40199b101294fac8c5c2d55ecc9e830eadf58c5646f71df41135b4375 WatchSource:0}: Error finding container 11bb68c40199b101294fac8c5c2d55ecc9e830eadf58c5646f71df41135b4375: Status 404 returned error can't find the container with id 11bb68c40199b101294fac8c5c2d55ecc9e830eadf58c5646f71df41135b4375 Nov 29 04:54:44 crc kubenswrapper[4799]: I1129 04:54:44.007809 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-2f96-account-create-update-4gdd6"] Nov 29 04:54:44 crc kubenswrapper[4799]: I1129 04:54:44.028732 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7451-account-create-update-r6m7w"] Nov 29 04:54:44 crc kubenswrapper[4799]: I1129 04:54:44.033168 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-gjbqp"] Nov 29 04:54:44 crc kubenswrapper[4799]: W1129 04:54:44.041125 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee16aa26_5b84_447d_8a02_7a4199451cc3.slice/crio-845895716b471f1e32abf85c08e803638003bc22be4ef40b776464174fe2a819 WatchSource:0}: Error finding container 845895716b471f1e32abf85c08e803638003bc22be4ef40b776464174fe2a819: Status 404 returned error can't find the container with id 845895716b471f1e32abf85c08e803638003bc22be4ef40b776464174fe2a819 Nov 29 04:54:44 crc kubenswrapper[4799]: I1129 04:54:44.223930 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-665zk" event={"ID":"ac99f10b-e4cb-4847-a60b-69306ff5e8ca","Type":"ContainerStarted","Data":"11bb68c40199b101294fac8c5c2d55ecc9e830eadf58c5646f71df41135b4375"} Nov 29 04:54:44 crc kubenswrapper[4799]: I1129 04:54:44.235503 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-88jbn" event={"ID":"c77dc962-f9a7-40dc-aeb6-d07e9c1be588","Type":"ContainerStarted","Data":"b523f31635771860b61f33361291c7fdc60502a4f3a198d87bc60d8298165fdc"} Nov 29 04:54:44 crc kubenswrapper[4799]: I1129 04:54:44.239280 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7451-account-create-update-r6m7w" event={"ID":"ee16aa26-5b84-447d-8a02-7a4199451cc3","Type":"ContainerStarted","Data":"845895716b471f1e32abf85c08e803638003bc22be4ef40b776464174fe2a819"} Nov 29 04:54:44 crc kubenswrapper[4799]: I1129 04:54:44.254149 4799 generic.go:334] "Generic (PLEG): container finished" podID="1a762825-986d-4691-88bb-da6580cf83b1" containerID="1c498b3d5cde10e3f6ad0de28e3513ceb35283bd6a48265e31da17bb6ab8d6de" exitCode=0 Nov 29 04:54:44 crc kubenswrapper[4799]: I1129 04:54:44.254407 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-9vlv8" event={"ID":"1a762825-986d-4691-88bb-da6580cf83b1","Type":"ContainerDied","Data":"1c498b3d5cde10e3f6ad0de28e3513ceb35283bd6a48265e31da17bb6ab8d6de"} Nov 29 04:54:44 crc kubenswrapper[4799]: I1129 04:54:44.254439 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-9vlv8" event={"ID":"1a762825-986d-4691-88bb-da6580cf83b1","Type":"ContainerStarted","Data":"f4902acb6f541b43569ee0d46a21153657f536232e4a91fcc3fb0f55f4791b96"} Nov 29 04:54:44 crc kubenswrapper[4799]: I1129 04:54:44.261047 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-2f96-account-create-update-4gdd6" event={"ID":"29be7ae1-328d-4e71-9796-f054b8f4686f","Type":"ContainerStarted","Data":"a6c7447cd9edc125e3449a93a26eb25146e25208f6d72b0841c2e63184921ca2"} Nov 29 04:54:44 crc kubenswrapper[4799]: I1129 04:54:44.263220 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-tv4bk" event={"ID":"fc7d764a-f94f-4f3a-bf98-643a323f7cc4","Type":"ContainerStarted","Data":"cd6bea142029b1fe43179aa046f1515ea6fafd02ee0f666c3c9647121a90d3ee"} Nov 29 04:54:44 crc kubenswrapper[4799]: I1129 04:54:44.263248 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-tv4bk" event={"ID":"fc7d764a-f94f-4f3a-bf98-643a323f7cc4","Type":"ContainerStarted","Data":"957d7bdec552d77a16d027232ac19c3bd61b08b6cf5ac45e001ef338191ea537"} Nov 29 04:54:44 crc kubenswrapper[4799]: I1129 04:54:44.272151 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-88jbn" podStartSLOduration=4.272282699 podStartE2EDuration="19.27211843s" podCreationTimestamp="2025-11-29 04:54:25 +0000 UTC" firstStartedPulling="2025-11-29 04:54:27.087651867 +0000 UTC m=+942.730582267" lastFinishedPulling="2025-11-29 04:54:42.087487598 +0000 UTC m=+957.730417998" observedRunningTime="2025-11-29 04:54:44.265901272 +0000 UTC m=+959.908831672" watchObservedRunningTime="2025-11-29 04:54:44.27211843 +0000 UTC m=+959.915048830" Nov 29 04:54:44 crc kubenswrapper[4799]: I1129 04:54:44.276239 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-gjbqp" event={"ID":"b5832499-91f5-48ca-9502-0931ae1ece9e","Type":"ContainerStarted","Data":"8186aad108b0efd0bf9ef72bfee6a66a000a0d703e5f4476128023ac71d36946"} Nov 29 04:54:44 crc kubenswrapper[4799]: I1129 04:54:44.282706 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-7da7-account-create-update-wr6xq" event={"ID":"b5536f91-0fec-4581-8a6d-18b365e7978b","Type":"ContainerStarted","Data":"52bf723d9bd89f6715e459c40f7837159a5187808f411c2d919780b16b3e05bc"} Nov 29 04:54:44 crc kubenswrapper[4799]: I1129 04:54:44.677234 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36215eda-2af7-4388-992f-b31807c9d423" path="/var/lib/kubelet/pods/36215eda-2af7-4388-992f-b31807c9d423/volumes" Nov 29 04:54:44 crc kubenswrapper[4799]: I1129 04:54:44.678891 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="631d5b8e-6703-4cee-96d7-4d014e38112e" path="/var/lib/kubelet/pods/631d5b8e-6703-4cee-96d7-4d014e38112e/volumes" Nov 29 04:54:45 crc kubenswrapper[4799]: I1129 04:54:45.327093 4799 generic.go:334] "Generic (PLEG): container finished" podID="ac99f10b-e4cb-4847-a60b-69306ff5e8ca" containerID="8cfbc8ad7e190a8189c171f5196c34588d6cf9c2acc922e7de3910d7bdcbfb1d" exitCode=0 Nov 29 04:54:45 crc kubenswrapper[4799]: I1129 04:54:45.327213 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-665zk" event={"ID":"ac99f10b-e4cb-4847-a60b-69306ff5e8ca","Type":"ContainerDied","Data":"8cfbc8ad7e190a8189c171f5196c34588d6cf9c2acc922e7de3910d7bdcbfb1d"} Nov 29 04:54:45 crc kubenswrapper[4799]: I1129 04:54:45.329892 4799 generic.go:334] "Generic (PLEG): container finished" podID="29be7ae1-328d-4e71-9796-f054b8f4686f" containerID="14a27c445ae2fa52c9d8c0fbfaee89d0939bf5cbaebbd9c13eb82c428f831151" exitCode=0 Nov 29 04:54:45 crc kubenswrapper[4799]: I1129 04:54:45.329958 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-2f96-account-create-update-4gdd6" event={"ID":"29be7ae1-328d-4e71-9796-f054b8f4686f","Type":"ContainerDied","Data":"14a27c445ae2fa52c9d8c0fbfaee89d0939bf5cbaebbd9c13eb82c428f831151"} Nov 29 04:54:45 crc kubenswrapper[4799]: I1129 04:54:45.331673 4799 generic.go:334] "Generic (PLEG): container finished" podID="ee16aa26-5b84-447d-8a02-7a4199451cc3" containerID="95a57181803d7c07e2e0c77db52741f06823e6d8f0fb7edf27b6731b7aefdbcf" exitCode=0 Nov 29 04:54:45 crc kubenswrapper[4799]: I1129 04:54:45.331734 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7451-account-create-update-r6m7w" event={"ID":"ee16aa26-5b84-447d-8a02-7a4199451cc3","Type":"ContainerDied","Data":"95a57181803d7c07e2e0c77db52741f06823e6d8f0fb7edf27b6731b7aefdbcf"} Nov 29 04:54:45 crc kubenswrapper[4799]: I1129 04:54:45.336507 4799 generic.go:334] "Generic (PLEG): container finished" podID="fc7d764a-f94f-4f3a-bf98-643a323f7cc4" containerID="cd6bea142029b1fe43179aa046f1515ea6fafd02ee0f666c3c9647121a90d3ee" exitCode=0 Nov 29 04:54:45 crc kubenswrapper[4799]: I1129 04:54:45.336581 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-tv4bk" event={"ID":"fc7d764a-f94f-4f3a-bf98-643a323f7cc4","Type":"ContainerDied","Data":"cd6bea142029b1fe43179aa046f1515ea6fafd02ee0f666c3c9647121a90d3ee"} Nov 29 04:54:45 crc kubenswrapper[4799]: I1129 04:54:45.338927 4799 generic.go:334] "Generic (PLEG): container finished" podID="b5536f91-0fec-4581-8a6d-18b365e7978b" containerID="c54c73645b3442b36048196436e145256b136c8e6ff2a4ee944d7ff41c3bdcb6" exitCode=0 Nov 29 04:54:45 crc kubenswrapper[4799]: I1129 04:54:45.339885 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-7da7-account-create-update-wr6xq" event={"ID":"b5536f91-0fec-4581-8a6d-18b365e7978b","Type":"ContainerDied","Data":"c54c73645b3442b36048196436e145256b136c8e6ff2a4ee944d7ff41c3bdcb6"} Nov 29 04:54:45 crc kubenswrapper[4799]: I1129 04:54:45.780346 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-tv4bk" Nov 29 04:54:45 crc kubenswrapper[4799]: I1129 04:54:45.787290 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-9vlv8" Nov 29 04:54:45 crc kubenswrapper[4799]: I1129 04:54:45.939495 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc7d764a-f94f-4f3a-bf98-643a323f7cc4-operator-scripts\") pod \"fc7d764a-f94f-4f3a-bf98-643a323f7cc4\" (UID: \"fc7d764a-f94f-4f3a-bf98-643a323f7cc4\") " Nov 29 04:54:45 crc kubenswrapper[4799]: I1129 04:54:45.939558 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mq5xs\" (UniqueName: \"kubernetes.io/projected/fc7d764a-f94f-4f3a-bf98-643a323f7cc4-kube-api-access-mq5xs\") pod \"fc7d764a-f94f-4f3a-bf98-643a323f7cc4\" (UID: \"fc7d764a-f94f-4f3a-bf98-643a323f7cc4\") " Nov 29 04:54:45 crc kubenswrapper[4799]: I1129 04:54:45.939587 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1a762825-986d-4691-88bb-da6580cf83b1-operator-scripts\") pod \"1a762825-986d-4691-88bb-da6580cf83b1\" (UID: \"1a762825-986d-4691-88bb-da6580cf83b1\") " Nov 29 04:54:45 crc kubenswrapper[4799]: I1129 04:54:45.939809 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8kqq9\" (UniqueName: \"kubernetes.io/projected/1a762825-986d-4691-88bb-da6580cf83b1-kube-api-access-8kqq9\") pod \"1a762825-986d-4691-88bb-da6580cf83b1\" (UID: \"1a762825-986d-4691-88bb-da6580cf83b1\") " Nov 29 04:54:45 crc kubenswrapper[4799]: I1129 04:54:45.940648 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a762825-986d-4691-88bb-da6580cf83b1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1a762825-986d-4691-88bb-da6580cf83b1" (UID: "1a762825-986d-4691-88bb-da6580cf83b1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:54:45 crc kubenswrapper[4799]: I1129 04:54:45.940768 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc7d764a-f94f-4f3a-bf98-643a323f7cc4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fc7d764a-f94f-4f3a-bf98-643a323f7cc4" (UID: "fc7d764a-f94f-4f3a-bf98-643a323f7cc4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:54:45 crc kubenswrapper[4799]: I1129 04:54:45.962416 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc7d764a-f94f-4f3a-bf98-643a323f7cc4-kube-api-access-mq5xs" (OuterVolumeSpecName: "kube-api-access-mq5xs") pod "fc7d764a-f94f-4f3a-bf98-643a323f7cc4" (UID: "fc7d764a-f94f-4f3a-bf98-643a323f7cc4"). InnerVolumeSpecName "kube-api-access-mq5xs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:54:45 crc kubenswrapper[4799]: I1129 04:54:45.962725 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a762825-986d-4691-88bb-da6580cf83b1-kube-api-access-8kqq9" (OuterVolumeSpecName: "kube-api-access-8kqq9") pod "1a762825-986d-4691-88bb-da6580cf83b1" (UID: "1a762825-986d-4691-88bb-da6580cf83b1"). InnerVolumeSpecName "kube-api-access-8kqq9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:54:46 crc kubenswrapper[4799]: I1129 04:54:46.042444 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8kqq9\" (UniqueName: \"kubernetes.io/projected/1a762825-986d-4691-88bb-da6580cf83b1-kube-api-access-8kqq9\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:46 crc kubenswrapper[4799]: I1129 04:54:46.042486 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc7d764a-f94f-4f3a-bf98-643a323f7cc4-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:46 crc kubenswrapper[4799]: I1129 04:54:46.042502 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mq5xs\" (UniqueName: \"kubernetes.io/projected/fc7d764a-f94f-4f3a-bf98-643a323f7cc4-kube-api-access-mq5xs\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:46 crc kubenswrapper[4799]: I1129 04:54:46.042518 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1a762825-986d-4691-88bb-da6580cf83b1-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:46 crc kubenswrapper[4799]: I1129 04:54:46.348466 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-tv4bk" event={"ID":"fc7d764a-f94f-4f3a-bf98-643a323f7cc4","Type":"ContainerDied","Data":"957d7bdec552d77a16d027232ac19c3bd61b08b6cf5ac45e001ef338191ea537"} Nov 29 04:54:46 crc kubenswrapper[4799]: I1129 04:54:46.348556 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="957d7bdec552d77a16d027232ac19c3bd61b08b6cf5ac45e001ef338191ea537" Nov 29 04:54:46 crc kubenswrapper[4799]: I1129 04:54:46.348493 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-tv4bk" Nov 29 04:54:46 crc kubenswrapper[4799]: I1129 04:54:46.351183 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-9vlv8" event={"ID":"1a762825-986d-4691-88bb-da6580cf83b1","Type":"ContainerDied","Data":"f4902acb6f541b43569ee0d46a21153657f536232e4a91fcc3fb0f55f4791b96"} Nov 29 04:54:46 crc kubenswrapper[4799]: I1129 04:54:46.351240 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f4902acb6f541b43569ee0d46a21153657f536232e4a91fcc3fb0f55f4791b96" Nov 29 04:54:46 crc kubenswrapper[4799]: I1129 04:54:46.351298 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-9vlv8" Nov 29 04:54:46 crc kubenswrapper[4799]: I1129 04:54:46.714878 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7451-account-create-update-r6m7w" Nov 29 04:54:46 crc kubenswrapper[4799]: I1129 04:54:46.757923 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zrdtt\" (UniqueName: \"kubernetes.io/projected/ee16aa26-5b84-447d-8a02-7a4199451cc3-kube-api-access-zrdtt\") pod \"ee16aa26-5b84-447d-8a02-7a4199451cc3\" (UID: \"ee16aa26-5b84-447d-8a02-7a4199451cc3\") " Nov 29 04:54:46 crc kubenswrapper[4799]: I1129 04:54:46.757977 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee16aa26-5b84-447d-8a02-7a4199451cc3-operator-scripts\") pod \"ee16aa26-5b84-447d-8a02-7a4199451cc3\" (UID: \"ee16aa26-5b84-447d-8a02-7a4199451cc3\") " Nov 29 04:54:46 crc kubenswrapper[4799]: I1129 04:54:46.759589 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee16aa26-5b84-447d-8a02-7a4199451cc3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ee16aa26-5b84-447d-8a02-7a4199451cc3" (UID: "ee16aa26-5b84-447d-8a02-7a4199451cc3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:54:46 crc kubenswrapper[4799]: I1129 04:54:46.765191 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee16aa26-5b84-447d-8a02-7a4199451cc3-kube-api-access-zrdtt" (OuterVolumeSpecName: "kube-api-access-zrdtt") pod "ee16aa26-5b84-447d-8a02-7a4199451cc3" (UID: "ee16aa26-5b84-447d-8a02-7a4199451cc3"). InnerVolumeSpecName "kube-api-access-zrdtt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:54:46 crc kubenswrapper[4799]: I1129 04:54:46.859878 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zrdtt\" (UniqueName: \"kubernetes.io/projected/ee16aa26-5b84-447d-8a02-7a4199451cc3-kube-api-access-zrdtt\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:46 crc kubenswrapper[4799]: I1129 04:54:46.860334 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee16aa26-5b84-447d-8a02-7a4199451cc3-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:46 crc kubenswrapper[4799]: I1129 04:54:46.894340 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-7da7-account-create-update-wr6xq" Nov 29 04:54:46 crc kubenswrapper[4799]: I1129 04:54:46.910207 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-2f96-account-create-update-4gdd6" Nov 29 04:54:46 crc kubenswrapper[4799]: I1129 04:54:46.912760 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-665zk" Nov 29 04:54:46 crc kubenswrapper[4799]: I1129 04:54:46.962466 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-89p6t\" (UniqueName: \"kubernetes.io/projected/29be7ae1-328d-4e71-9796-f054b8f4686f-kube-api-access-89p6t\") pod \"29be7ae1-328d-4e71-9796-f054b8f4686f\" (UID: \"29be7ae1-328d-4e71-9796-f054b8f4686f\") " Nov 29 04:54:46 crc kubenswrapper[4799]: I1129 04:54:46.962568 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mfjg8\" (UniqueName: \"kubernetes.io/projected/b5536f91-0fec-4581-8a6d-18b365e7978b-kube-api-access-mfjg8\") pod \"b5536f91-0fec-4581-8a6d-18b365e7978b\" (UID: \"b5536f91-0fec-4581-8a6d-18b365e7978b\") " Nov 29 04:54:46 crc kubenswrapper[4799]: I1129 04:54:46.962598 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fbpr5\" (UniqueName: \"kubernetes.io/projected/ac99f10b-e4cb-4847-a60b-69306ff5e8ca-kube-api-access-fbpr5\") pod \"ac99f10b-e4cb-4847-a60b-69306ff5e8ca\" (UID: \"ac99f10b-e4cb-4847-a60b-69306ff5e8ca\") " Nov 29 04:54:46 crc kubenswrapper[4799]: I1129 04:54:46.962631 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/29be7ae1-328d-4e71-9796-f054b8f4686f-operator-scripts\") pod \"29be7ae1-328d-4e71-9796-f054b8f4686f\" (UID: \"29be7ae1-328d-4e71-9796-f054b8f4686f\") " Nov 29 04:54:46 crc kubenswrapper[4799]: I1129 04:54:46.962839 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ac99f10b-e4cb-4847-a60b-69306ff5e8ca-operator-scripts\") pod \"ac99f10b-e4cb-4847-a60b-69306ff5e8ca\" (UID: \"ac99f10b-e4cb-4847-a60b-69306ff5e8ca\") " Nov 29 04:54:46 crc kubenswrapper[4799]: I1129 04:54:46.962915 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5536f91-0fec-4581-8a6d-18b365e7978b-operator-scripts\") pod \"b5536f91-0fec-4581-8a6d-18b365e7978b\" (UID: \"b5536f91-0fec-4581-8a6d-18b365e7978b\") " Nov 29 04:54:46 crc kubenswrapper[4799]: I1129 04:54:46.963412 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/29be7ae1-328d-4e71-9796-f054b8f4686f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "29be7ae1-328d-4e71-9796-f054b8f4686f" (UID: "29be7ae1-328d-4e71-9796-f054b8f4686f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:54:46 crc kubenswrapper[4799]: I1129 04:54:46.963490 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5536f91-0fec-4581-8a6d-18b365e7978b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b5536f91-0fec-4581-8a6d-18b365e7978b" (UID: "b5536f91-0fec-4581-8a6d-18b365e7978b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:54:46 crc kubenswrapper[4799]: I1129 04:54:46.963901 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac99f10b-e4cb-4847-a60b-69306ff5e8ca-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ac99f10b-e4cb-4847-a60b-69306ff5e8ca" (UID: "ac99f10b-e4cb-4847-a60b-69306ff5e8ca"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:54:46 crc kubenswrapper[4799]: I1129 04:54:46.972011 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac99f10b-e4cb-4847-a60b-69306ff5e8ca-kube-api-access-fbpr5" (OuterVolumeSpecName: "kube-api-access-fbpr5") pod "ac99f10b-e4cb-4847-a60b-69306ff5e8ca" (UID: "ac99f10b-e4cb-4847-a60b-69306ff5e8ca"). InnerVolumeSpecName "kube-api-access-fbpr5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:54:46 crc kubenswrapper[4799]: I1129 04:54:46.972329 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5536f91-0fec-4581-8a6d-18b365e7978b-kube-api-access-mfjg8" (OuterVolumeSpecName: "kube-api-access-mfjg8") pod "b5536f91-0fec-4581-8a6d-18b365e7978b" (UID: "b5536f91-0fec-4581-8a6d-18b365e7978b"). InnerVolumeSpecName "kube-api-access-mfjg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:54:46 crc kubenswrapper[4799]: I1129 04:54:46.973412 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29be7ae1-328d-4e71-9796-f054b8f4686f-kube-api-access-89p6t" (OuterVolumeSpecName: "kube-api-access-89p6t") pod "29be7ae1-328d-4e71-9796-f054b8f4686f" (UID: "29be7ae1-328d-4e71-9796-f054b8f4686f"). InnerVolumeSpecName "kube-api-access-89p6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:54:47 crc kubenswrapper[4799]: I1129 04:54:47.065861 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5536f91-0fec-4581-8a6d-18b365e7978b-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:47 crc kubenswrapper[4799]: I1129 04:54:47.065903 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-89p6t\" (UniqueName: \"kubernetes.io/projected/29be7ae1-328d-4e71-9796-f054b8f4686f-kube-api-access-89p6t\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:47 crc kubenswrapper[4799]: I1129 04:54:47.065916 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mfjg8\" (UniqueName: \"kubernetes.io/projected/b5536f91-0fec-4581-8a6d-18b365e7978b-kube-api-access-mfjg8\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:47 crc kubenswrapper[4799]: I1129 04:54:47.065926 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fbpr5\" (UniqueName: \"kubernetes.io/projected/ac99f10b-e4cb-4847-a60b-69306ff5e8ca-kube-api-access-fbpr5\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:47 crc kubenswrapper[4799]: I1129 04:54:47.065936 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/29be7ae1-328d-4e71-9796-f054b8f4686f-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:47 crc kubenswrapper[4799]: I1129 04:54:47.065944 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ac99f10b-e4cb-4847-a60b-69306ff5e8ca-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:47 crc kubenswrapper[4799]: I1129 04:54:47.363065 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-7da7-account-create-update-wr6xq" event={"ID":"b5536f91-0fec-4581-8a6d-18b365e7978b","Type":"ContainerDied","Data":"52bf723d9bd89f6715e459c40f7837159a5187808f411c2d919780b16b3e05bc"} Nov 29 04:54:47 crc kubenswrapper[4799]: I1129 04:54:47.363994 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="52bf723d9bd89f6715e459c40f7837159a5187808f411c2d919780b16b3e05bc" Nov 29 04:54:47 crc kubenswrapper[4799]: I1129 04:54:47.363148 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-7da7-account-create-update-wr6xq" Nov 29 04:54:47 crc kubenswrapper[4799]: I1129 04:54:47.366155 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-665zk" event={"ID":"ac99f10b-e4cb-4847-a60b-69306ff5e8ca","Type":"ContainerDied","Data":"11bb68c40199b101294fac8c5c2d55ecc9e830eadf58c5646f71df41135b4375"} Nov 29 04:54:47 crc kubenswrapper[4799]: I1129 04:54:47.366235 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="11bb68c40199b101294fac8c5c2d55ecc9e830eadf58c5646f71df41135b4375" Nov 29 04:54:47 crc kubenswrapper[4799]: I1129 04:54:47.366255 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-665zk" Nov 29 04:54:47 crc kubenswrapper[4799]: I1129 04:54:47.371070 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-2f96-account-create-update-4gdd6" event={"ID":"29be7ae1-328d-4e71-9796-f054b8f4686f","Type":"ContainerDied","Data":"a6c7447cd9edc125e3449a93a26eb25146e25208f6d72b0841c2e63184921ca2"} Nov 29 04:54:47 crc kubenswrapper[4799]: I1129 04:54:47.371146 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6c7447cd9edc125e3449a93a26eb25146e25208f6d72b0841c2e63184921ca2" Nov 29 04:54:47 crc kubenswrapper[4799]: I1129 04:54:47.371084 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-2f96-account-create-update-4gdd6" Nov 29 04:54:47 crc kubenswrapper[4799]: I1129 04:54:47.377884 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7451-account-create-update-r6m7w" event={"ID":"ee16aa26-5b84-447d-8a02-7a4199451cc3","Type":"ContainerDied","Data":"845895716b471f1e32abf85c08e803638003bc22be4ef40b776464174fe2a819"} Nov 29 04:54:47 crc kubenswrapper[4799]: I1129 04:54:47.377940 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="845895716b471f1e32abf85c08e803638003bc22be4ef40b776464174fe2a819" Nov 29 04:54:47 crc kubenswrapper[4799]: I1129 04:54:47.377938 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7451-account-create-update-r6m7w" Nov 29 04:54:51 crc kubenswrapper[4799]: I1129 04:54:51.434002 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-gjbqp" event={"ID":"b5832499-91f5-48ca-9502-0931ae1ece9e","Type":"ContainerStarted","Data":"bd94990dc745ca3eae2695b4842f3b3fb903f87a853ba4e52e1a83fb34911e40"} Nov 29 04:54:51 crc kubenswrapper[4799]: I1129 04:54:51.452843 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-gjbqp" podStartSLOduration=2.43223257 podStartE2EDuration="9.452816745s" podCreationTimestamp="2025-11-29 04:54:42 +0000 UTC" firstStartedPulling="2025-11-29 04:54:44.099519712 +0000 UTC m=+959.742450112" lastFinishedPulling="2025-11-29 04:54:51.120103887 +0000 UTC m=+966.763034287" observedRunningTime="2025-11-29 04:54:51.447190102 +0000 UTC m=+967.090120522" watchObservedRunningTime="2025-11-29 04:54:51.452816745 +0000 UTC m=+967.095747145" Nov 29 04:54:53 crc kubenswrapper[4799]: I1129 04:54:53.458353 4799 generic.go:334] "Generic (PLEG): container finished" podID="c77dc962-f9a7-40dc-aeb6-d07e9c1be588" containerID="b523f31635771860b61f33361291c7fdc60502a4f3a198d87bc60d8298165fdc" exitCode=0 Nov 29 04:54:53 crc kubenswrapper[4799]: I1129 04:54:53.458447 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-88jbn" event={"ID":"c77dc962-f9a7-40dc-aeb6-d07e9c1be588","Type":"ContainerDied","Data":"b523f31635771860b61f33361291c7fdc60502a4f3a198d87bc60d8298165fdc"} Nov 29 04:54:54 crc kubenswrapper[4799]: I1129 04:54:54.470877 4799 generic.go:334] "Generic (PLEG): container finished" podID="b5832499-91f5-48ca-9502-0931ae1ece9e" containerID="bd94990dc745ca3eae2695b4842f3b3fb903f87a853ba4e52e1a83fb34911e40" exitCode=0 Nov 29 04:54:54 crc kubenswrapper[4799]: I1129 04:54:54.470975 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-gjbqp" event={"ID":"b5832499-91f5-48ca-9502-0931ae1ece9e","Type":"ContainerDied","Data":"bd94990dc745ca3eae2695b4842f3b3fb903f87a853ba4e52e1a83fb34911e40"} Nov 29 04:54:55 crc kubenswrapper[4799]: I1129 04:54:55.315220 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-88jbn" Nov 29 04:54:55 crc kubenswrapper[4799]: I1129 04:54:55.482084 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-88jbn" Nov 29 04:54:55 crc kubenswrapper[4799]: I1129 04:54:55.482075 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-88jbn" event={"ID":"c77dc962-f9a7-40dc-aeb6-d07e9c1be588","Type":"ContainerDied","Data":"10523f9715769be575965752a09ec60026d7d25d5a259f7484c0d2a5b591c4e6"} Nov 29 04:54:55 crc kubenswrapper[4799]: I1129 04:54:55.482158 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="10523f9715769be575965752a09ec60026d7d25d5a259f7484c0d2a5b591c4e6" Nov 29 04:54:55 crc kubenswrapper[4799]: I1129 04:54:55.493513 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c77dc962-f9a7-40dc-aeb6-d07e9c1be588-combined-ca-bundle\") pod \"c77dc962-f9a7-40dc-aeb6-d07e9c1be588\" (UID: \"c77dc962-f9a7-40dc-aeb6-d07e9c1be588\") " Nov 29 04:54:55 crc kubenswrapper[4799]: I1129 04:54:55.493584 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d8sfw\" (UniqueName: \"kubernetes.io/projected/c77dc962-f9a7-40dc-aeb6-d07e9c1be588-kube-api-access-d8sfw\") pod \"c77dc962-f9a7-40dc-aeb6-d07e9c1be588\" (UID: \"c77dc962-f9a7-40dc-aeb6-d07e9c1be588\") " Nov 29 04:54:55 crc kubenswrapper[4799]: I1129 04:54:55.493742 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c77dc962-f9a7-40dc-aeb6-d07e9c1be588-db-sync-config-data\") pod \"c77dc962-f9a7-40dc-aeb6-d07e9c1be588\" (UID: \"c77dc962-f9a7-40dc-aeb6-d07e9c1be588\") " Nov 29 04:54:55 crc kubenswrapper[4799]: I1129 04:54:55.493820 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c77dc962-f9a7-40dc-aeb6-d07e9c1be588-config-data\") pod \"c77dc962-f9a7-40dc-aeb6-d07e9c1be588\" (UID: \"c77dc962-f9a7-40dc-aeb6-d07e9c1be588\") " Nov 29 04:54:55 crc kubenswrapper[4799]: I1129 04:54:55.511529 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c77dc962-f9a7-40dc-aeb6-d07e9c1be588-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "c77dc962-f9a7-40dc-aeb6-d07e9c1be588" (UID: "c77dc962-f9a7-40dc-aeb6-d07e9c1be588"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:54:55 crc kubenswrapper[4799]: I1129 04:54:55.516664 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c77dc962-f9a7-40dc-aeb6-d07e9c1be588-kube-api-access-d8sfw" (OuterVolumeSpecName: "kube-api-access-d8sfw") pod "c77dc962-f9a7-40dc-aeb6-d07e9c1be588" (UID: "c77dc962-f9a7-40dc-aeb6-d07e9c1be588"). InnerVolumeSpecName "kube-api-access-d8sfw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:54:55 crc kubenswrapper[4799]: I1129 04:54:55.531068 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c77dc962-f9a7-40dc-aeb6-d07e9c1be588-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c77dc962-f9a7-40dc-aeb6-d07e9c1be588" (UID: "c77dc962-f9a7-40dc-aeb6-d07e9c1be588"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:54:55 crc kubenswrapper[4799]: I1129 04:54:55.564381 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c77dc962-f9a7-40dc-aeb6-d07e9c1be588-config-data" (OuterVolumeSpecName: "config-data") pod "c77dc962-f9a7-40dc-aeb6-d07e9c1be588" (UID: "c77dc962-f9a7-40dc-aeb6-d07e9c1be588"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:54:55 crc kubenswrapper[4799]: I1129 04:54:55.595845 4799 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c77dc962-f9a7-40dc-aeb6-d07e9c1be588-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:55 crc kubenswrapper[4799]: I1129 04:54:55.595893 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c77dc962-f9a7-40dc-aeb6-d07e9c1be588-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:55 crc kubenswrapper[4799]: I1129 04:54:55.595906 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c77dc962-f9a7-40dc-aeb6-d07e9c1be588-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:55 crc kubenswrapper[4799]: I1129 04:54:55.595918 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d8sfw\" (UniqueName: \"kubernetes.io/projected/c77dc962-f9a7-40dc-aeb6-d07e9c1be588-kube-api-access-d8sfw\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:55 crc kubenswrapper[4799]: I1129 04:54:55.745723 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-gjbqp" Nov 29 04:54:55 crc kubenswrapper[4799]: I1129 04:54:55.891852 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-54f9b7b8d9-8sxd6"] Nov 29 04:54:55 crc kubenswrapper[4799]: E1129 04:54:55.892530 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c77dc962-f9a7-40dc-aeb6-d07e9c1be588" containerName="glance-db-sync" Nov 29 04:54:55 crc kubenswrapper[4799]: I1129 04:54:55.892548 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="c77dc962-f9a7-40dc-aeb6-d07e9c1be588" containerName="glance-db-sync" Nov 29 04:54:55 crc kubenswrapper[4799]: E1129 04:54:55.892572 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29be7ae1-328d-4e71-9796-f054b8f4686f" containerName="mariadb-account-create-update" Nov 29 04:54:55 crc kubenswrapper[4799]: I1129 04:54:55.892579 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="29be7ae1-328d-4e71-9796-f054b8f4686f" containerName="mariadb-account-create-update" Nov 29 04:54:55 crc kubenswrapper[4799]: E1129 04:54:55.892590 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5536f91-0fec-4581-8a6d-18b365e7978b" containerName="mariadb-account-create-update" Nov 29 04:54:55 crc kubenswrapper[4799]: I1129 04:54:55.892596 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5536f91-0fec-4581-8a6d-18b365e7978b" containerName="mariadb-account-create-update" Nov 29 04:54:55 crc kubenswrapper[4799]: E1129 04:54:55.892612 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee16aa26-5b84-447d-8a02-7a4199451cc3" containerName="mariadb-account-create-update" Nov 29 04:54:55 crc kubenswrapper[4799]: I1129 04:54:55.892618 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee16aa26-5b84-447d-8a02-7a4199451cc3" containerName="mariadb-account-create-update" Nov 29 04:54:55 crc kubenswrapper[4799]: E1129 04:54:55.892629 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a762825-986d-4691-88bb-da6580cf83b1" containerName="mariadb-database-create" Nov 29 04:54:55 crc kubenswrapper[4799]: I1129 04:54:55.892635 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a762825-986d-4691-88bb-da6580cf83b1" containerName="mariadb-database-create" Nov 29 04:54:55 crc kubenswrapper[4799]: E1129 04:54:55.892642 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac99f10b-e4cb-4847-a60b-69306ff5e8ca" containerName="mariadb-database-create" Nov 29 04:54:55 crc kubenswrapper[4799]: I1129 04:54:55.892649 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac99f10b-e4cb-4847-a60b-69306ff5e8ca" containerName="mariadb-database-create" Nov 29 04:54:55 crc kubenswrapper[4799]: E1129 04:54:55.892659 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5832499-91f5-48ca-9502-0931ae1ece9e" containerName="keystone-db-sync" Nov 29 04:54:55 crc kubenswrapper[4799]: I1129 04:54:55.892667 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5832499-91f5-48ca-9502-0931ae1ece9e" containerName="keystone-db-sync" Nov 29 04:54:55 crc kubenswrapper[4799]: E1129 04:54:55.892677 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc7d764a-f94f-4f3a-bf98-643a323f7cc4" containerName="mariadb-database-create" Nov 29 04:54:55 crc kubenswrapper[4799]: I1129 04:54:55.892685 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc7d764a-f94f-4f3a-bf98-643a323f7cc4" containerName="mariadb-database-create" Nov 29 04:54:55 crc kubenswrapper[4799]: I1129 04:54:55.892898 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5832499-91f5-48ca-9502-0931ae1ece9e" containerName="keystone-db-sync" Nov 29 04:54:55 crc kubenswrapper[4799]: I1129 04:54:55.892912 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="29be7ae1-328d-4e71-9796-f054b8f4686f" containerName="mariadb-account-create-update" Nov 29 04:54:55 crc kubenswrapper[4799]: I1129 04:54:55.892924 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a762825-986d-4691-88bb-da6580cf83b1" containerName="mariadb-database-create" Nov 29 04:54:55 crc kubenswrapper[4799]: I1129 04:54:55.892931 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5536f91-0fec-4581-8a6d-18b365e7978b" containerName="mariadb-account-create-update" Nov 29 04:54:55 crc kubenswrapper[4799]: I1129 04:54:55.892942 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee16aa26-5b84-447d-8a02-7a4199451cc3" containerName="mariadb-account-create-update" Nov 29 04:54:55 crc kubenswrapper[4799]: I1129 04:54:55.892951 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac99f10b-e4cb-4847-a60b-69306ff5e8ca" containerName="mariadb-database-create" Nov 29 04:54:55 crc kubenswrapper[4799]: I1129 04:54:55.892958 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="c77dc962-f9a7-40dc-aeb6-d07e9c1be588" containerName="glance-db-sync" Nov 29 04:54:55 crc kubenswrapper[4799]: I1129 04:54:55.892966 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc7d764a-f94f-4f3a-bf98-643a323f7cc4" containerName="mariadb-database-create" Nov 29 04:54:55 crc kubenswrapper[4799]: I1129 04:54:55.896406 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54f9b7b8d9-8sxd6" Nov 29 04:54:55 crc kubenswrapper[4799]: I1129 04:54:55.906464 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5832499-91f5-48ca-9502-0931ae1ece9e-config-data\") pod \"b5832499-91f5-48ca-9502-0931ae1ece9e\" (UID: \"b5832499-91f5-48ca-9502-0931ae1ece9e\") " Nov 29 04:54:55 crc kubenswrapper[4799]: I1129 04:54:55.906733 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ppp45\" (UniqueName: \"kubernetes.io/projected/b5832499-91f5-48ca-9502-0931ae1ece9e-kube-api-access-ppp45\") pod \"b5832499-91f5-48ca-9502-0931ae1ece9e\" (UID: \"b5832499-91f5-48ca-9502-0931ae1ece9e\") " Nov 29 04:54:55 crc kubenswrapper[4799]: I1129 04:54:55.906853 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5832499-91f5-48ca-9502-0931ae1ece9e-combined-ca-bundle\") pod \"b5832499-91f5-48ca-9502-0931ae1ece9e\" (UID: \"b5832499-91f5-48ca-9502-0931ae1ece9e\") " Nov 29 04:54:55 crc kubenswrapper[4799]: I1129 04:54:55.933665 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54f9b7b8d9-8sxd6"] Nov 29 04:54:55 crc kubenswrapper[4799]: I1129 04:54:55.937735 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5832499-91f5-48ca-9502-0931ae1ece9e-kube-api-access-ppp45" (OuterVolumeSpecName: "kube-api-access-ppp45") pod "b5832499-91f5-48ca-9502-0931ae1ece9e" (UID: "b5832499-91f5-48ca-9502-0931ae1ece9e"). InnerVolumeSpecName "kube-api-access-ppp45". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:54:55 crc kubenswrapper[4799]: I1129 04:54:55.940960 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5832499-91f5-48ca-9502-0931ae1ece9e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b5832499-91f5-48ca-9502-0931ae1ece9e" (UID: "b5832499-91f5-48ca-9502-0931ae1ece9e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:54:55 crc kubenswrapper[4799]: I1129 04:54:55.978949 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5832499-91f5-48ca-9502-0931ae1ece9e-config-data" (OuterVolumeSpecName: "config-data") pod "b5832499-91f5-48ca-9502-0931ae1ece9e" (UID: "b5832499-91f5-48ca-9502-0931ae1ece9e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:54:56 crc kubenswrapper[4799]: I1129 04:54:56.009371 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/64328f76-389c-43d8-9f70-e6888983268e-ovsdbserver-nb\") pod \"dnsmasq-dns-54f9b7b8d9-8sxd6\" (UID: \"64328f76-389c-43d8-9f70-e6888983268e\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-8sxd6" Nov 29 04:54:56 crc kubenswrapper[4799]: I1129 04:54:56.009416 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbj6n\" (UniqueName: \"kubernetes.io/projected/64328f76-389c-43d8-9f70-e6888983268e-kube-api-access-hbj6n\") pod \"dnsmasq-dns-54f9b7b8d9-8sxd6\" (UID: \"64328f76-389c-43d8-9f70-e6888983268e\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-8sxd6" Nov 29 04:54:56 crc kubenswrapper[4799]: I1129 04:54:56.009463 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/64328f76-389c-43d8-9f70-e6888983268e-ovsdbserver-sb\") pod \"dnsmasq-dns-54f9b7b8d9-8sxd6\" (UID: \"64328f76-389c-43d8-9f70-e6888983268e\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-8sxd6" Nov 29 04:54:56 crc kubenswrapper[4799]: I1129 04:54:56.009767 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/64328f76-389c-43d8-9f70-e6888983268e-dns-svc\") pod \"dnsmasq-dns-54f9b7b8d9-8sxd6\" (UID: \"64328f76-389c-43d8-9f70-e6888983268e\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-8sxd6" Nov 29 04:54:56 crc kubenswrapper[4799]: I1129 04:54:56.009972 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64328f76-389c-43d8-9f70-e6888983268e-config\") pod \"dnsmasq-dns-54f9b7b8d9-8sxd6\" (UID: \"64328f76-389c-43d8-9f70-e6888983268e\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-8sxd6" Nov 29 04:54:56 crc kubenswrapper[4799]: I1129 04:54:56.010207 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5832499-91f5-48ca-9502-0931ae1ece9e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:56 crc kubenswrapper[4799]: I1129 04:54:56.010229 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5832499-91f5-48ca-9502-0931ae1ece9e-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:56 crc kubenswrapper[4799]: I1129 04:54:56.010242 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ppp45\" (UniqueName: \"kubernetes.io/projected/b5832499-91f5-48ca-9502-0931ae1ece9e-kube-api-access-ppp45\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:56 crc kubenswrapper[4799]: I1129 04:54:56.111496 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/64328f76-389c-43d8-9f70-e6888983268e-ovsdbserver-nb\") pod \"dnsmasq-dns-54f9b7b8d9-8sxd6\" (UID: \"64328f76-389c-43d8-9f70-e6888983268e\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-8sxd6" Nov 29 04:54:56 crc kubenswrapper[4799]: I1129 04:54:56.111560 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbj6n\" (UniqueName: \"kubernetes.io/projected/64328f76-389c-43d8-9f70-e6888983268e-kube-api-access-hbj6n\") pod \"dnsmasq-dns-54f9b7b8d9-8sxd6\" (UID: \"64328f76-389c-43d8-9f70-e6888983268e\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-8sxd6" Nov 29 04:54:56 crc kubenswrapper[4799]: I1129 04:54:56.111617 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/64328f76-389c-43d8-9f70-e6888983268e-ovsdbserver-sb\") pod \"dnsmasq-dns-54f9b7b8d9-8sxd6\" (UID: \"64328f76-389c-43d8-9f70-e6888983268e\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-8sxd6" Nov 29 04:54:56 crc kubenswrapper[4799]: I1129 04:54:56.111663 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/64328f76-389c-43d8-9f70-e6888983268e-dns-svc\") pod \"dnsmasq-dns-54f9b7b8d9-8sxd6\" (UID: \"64328f76-389c-43d8-9f70-e6888983268e\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-8sxd6" Nov 29 04:54:56 crc kubenswrapper[4799]: I1129 04:54:56.111705 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64328f76-389c-43d8-9f70-e6888983268e-config\") pod \"dnsmasq-dns-54f9b7b8d9-8sxd6\" (UID: \"64328f76-389c-43d8-9f70-e6888983268e\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-8sxd6" Nov 29 04:54:56 crc kubenswrapper[4799]: I1129 04:54:56.112914 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/64328f76-389c-43d8-9f70-e6888983268e-ovsdbserver-nb\") pod \"dnsmasq-dns-54f9b7b8d9-8sxd6\" (UID: \"64328f76-389c-43d8-9f70-e6888983268e\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-8sxd6" Nov 29 04:54:56 crc kubenswrapper[4799]: I1129 04:54:56.113019 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/64328f76-389c-43d8-9f70-e6888983268e-dns-svc\") pod \"dnsmasq-dns-54f9b7b8d9-8sxd6\" (UID: \"64328f76-389c-43d8-9f70-e6888983268e\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-8sxd6" Nov 29 04:54:56 crc kubenswrapper[4799]: I1129 04:54:56.113255 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64328f76-389c-43d8-9f70-e6888983268e-config\") pod \"dnsmasq-dns-54f9b7b8d9-8sxd6\" (UID: \"64328f76-389c-43d8-9f70-e6888983268e\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-8sxd6" Nov 29 04:54:56 crc kubenswrapper[4799]: I1129 04:54:56.113555 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/64328f76-389c-43d8-9f70-e6888983268e-ovsdbserver-sb\") pod \"dnsmasq-dns-54f9b7b8d9-8sxd6\" (UID: \"64328f76-389c-43d8-9f70-e6888983268e\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-8sxd6" Nov 29 04:54:56 crc kubenswrapper[4799]: I1129 04:54:56.129252 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbj6n\" (UniqueName: \"kubernetes.io/projected/64328f76-389c-43d8-9f70-e6888983268e-kube-api-access-hbj6n\") pod \"dnsmasq-dns-54f9b7b8d9-8sxd6\" (UID: \"64328f76-389c-43d8-9f70-e6888983268e\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-8sxd6" Nov 29 04:54:56 crc kubenswrapper[4799]: I1129 04:54:56.217605 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54f9b7b8d9-8sxd6" Nov 29 04:54:56 crc kubenswrapper[4799]: I1129 04:54:56.535186 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-gjbqp" event={"ID":"b5832499-91f5-48ca-9502-0931ae1ece9e","Type":"ContainerDied","Data":"8186aad108b0efd0bf9ef72bfee6a66a000a0d703e5f4476128023ac71d36946"} Nov 29 04:54:56 crc kubenswrapper[4799]: I1129 04:54:56.535570 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8186aad108b0efd0bf9ef72bfee6a66a000a0d703e5f4476128023ac71d36946" Nov 29 04:54:56 crc kubenswrapper[4799]: I1129 04:54:56.535298 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-gjbqp" Nov 29 04:54:56 crc kubenswrapper[4799]: I1129 04:54:56.745454 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54f9b7b8d9-8sxd6"] Nov 29 04:54:56 crc kubenswrapper[4799]: I1129 04:54:56.791391 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54f9b7b8d9-8sxd6"] Nov 29 04:54:56 crc kubenswrapper[4799]: I1129 04:54:56.811619 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-sz8zj"] Nov 29 04:54:56 crc kubenswrapper[4799]: I1129 04:54:56.813700 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-sz8zj" Nov 29 04:54:56 crc kubenswrapper[4799]: I1129 04:54:56.816532 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 29 04:54:56 crc kubenswrapper[4799]: I1129 04:54:56.822546 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 29 04:54:56 crc kubenswrapper[4799]: I1129 04:54:56.822828 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-85vg6" Nov 29 04:54:56 crc kubenswrapper[4799]: I1129 04:54:56.822958 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 29 04:54:56 crc kubenswrapper[4799]: I1129 04:54:56.823128 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 29 04:54:56 crc kubenswrapper[4799]: I1129 04:54:56.849340 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6546db6db7-w7zk2"] Nov 29 04:54:56 crc kubenswrapper[4799]: I1129 04:54:56.851062 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6546db6db7-w7zk2" Nov 29 04:54:56 crc kubenswrapper[4799]: I1129 04:54:56.857270 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6546db6db7-w7zk2"] Nov 29 04:54:56 crc kubenswrapper[4799]: I1129 04:54:56.882953 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-sz8zj"] Nov 29 04:54:56 crc kubenswrapper[4799]: I1129 04:54:56.937324 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3e1b036a-a1f1-44ed-9823-b49199c5b4de-credential-keys\") pod \"keystone-bootstrap-sz8zj\" (UID: \"3e1b036a-a1f1-44ed-9823-b49199c5b4de\") " pod="openstack/keystone-bootstrap-sz8zj" Nov 29 04:54:56 crc kubenswrapper[4799]: I1129 04:54:56.937499 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbsbx\" (UniqueName: \"kubernetes.io/projected/2dabdbea-d929-4b73-85a4-b6a882967131-kube-api-access-pbsbx\") pod \"dnsmasq-dns-6546db6db7-w7zk2\" (UID: \"2dabdbea-d929-4b73-85a4-b6a882967131\") " pod="openstack/dnsmasq-dns-6546db6db7-w7zk2" Nov 29 04:54:56 crc kubenswrapper[4799]: I1129 04:54:56.937650 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2dabdbea-d929-4b73-85a4-b6a882967131-ovsdbserver-sb\") pod \"dnsmasq-dns-6546db6db7-w7zk2\" (UID: \"2dabdbea-d929-4b73-85a4-b6a882967131\") " pod="openstack/dnsmasq-dns-6546db6db7-w7zk2" Nov 29 04:54:56 crc kubenswrapper[4799]: I1129 04:54:56.937673 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e1b036a-a1f1-44ed-9823-b49199c5b4de-config-data\") pod \"keystone-bootstrap-sz8zj\" (UID: \"3e1b036a-a1f1-44ed-9823-b49199c5b4de\") " pod="openstack/keystone-bootstrap-sz8zj" Nov 29 04:54:56 crc kubenswrapper[4799]: I1129 04:54:56.937692 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3e1b036a-a1f1-44ed-9823-b49199c5b4de-fernet-keys\") pod \"keystone-bootstrap-sz8zj\" (UID: \"3e1b036a-a1f1-44ed-9823-b49199c5b4de\") " pod="openstack/keystone-bootstrap-sz8zj" Nov 29 04:54:56 crc kubenswrapper[4799]: I1129 04:54:56.937729 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2dabdbea-d929-4b73-85a4-b6a882967131-dns-svc\") pod \"dnsmasq-dns-6546db6db7-w7zk2\" (UID: \"2dabdbea-d929-4b73-85a4-b6a882967131\") " pod="openstack/dnsmasq-dns-6546db6db7-w7zk2" Nov 29 04:54:56 crc kubenswrapper[4799]: I1129 04:54:56.937821 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e1b036a-a1f1-44ed-9823-b49199c5b4de-scripts\") pod \"keystone-bootstrap-sz8zj\" (UID: \"3e1b036a-a1f1-44ed-9823-b49199c5b4de\") " pod="openstack/keystone-bootstrap-sz8zj" Nov 29 04:54:56 crc kubenswrapper[4799]: I1129 04:54:56.937882 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kh29k\" (UniqueName: \"kubernetes.io/projected/3e1b036a-a1f1-44ed-9823-b49199c5b4de-kube-api-access-kh29k\") pod \"keystone-bootstrap-sz8zj\" (UID: \"3e1b036a-a1f1-44ed-9823-b49199c5b4de\") " pod="openstack/keystone-bootstrap-sz8zj" Nov 29 04:54:56 crc kubenswrapper[4799]: I1129 04:54:56.937914 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2dabdbea-d929-4b73-85a4-b6a882967131-ovsdbserver-nb\") pod \"dnsmasq-dns-6546db6db7-w7zk2\" (UID: \"2dabdbea-d929-4b73-85a4-b6a882967131\") " pod="openstack/dnsmasq-dns-6546db6db7-w7zk2" Nov 29 04:54:56 crc kubenswrapper[4799]: I1129 04:54:56.937972 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2dabdbea-d929-4b73-85a4-b6a882967131-config\") pod \"dnsmasq-dns-6546db6db7-w7zk2\" (UID: \"2dabdbea-d929-4b73-85a4-b6a882967131\") " pod="openstack/dnsmasq-dns-6546db6db7-w7zk2" Nov 29 04:54:56 crc kubenswrapper[4799]: I1129 04:54:56.938065 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e1b036a-a1f1-44ed-9823-b49199c5b4de-combined-ca-bundle\") pod \"keystone-bootstrap-sz8zj\" (UID: \"3e1b036a-a1f1-44ed-9823-b49199c5b4de\") " pod="openstack/keystone-bootstrap-sz8zj" Nov 29 04:54:56 crc kubenswrapper[4799]: I1129 04:54:56.993378 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-btd8j"] Nov 29 04:54:56 crc kubenswrapper[4799]: I1129 04:54:56.995105 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-btd8j" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.001277 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-hvxx6" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.001620 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.001749 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.025442 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-btd8j"] Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.041077 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3e1b036a-a1f1-44ed-9823-b49199c5b4de-credential-keys\") pod \"keystone-bootstrap-sz8zj\" (UID: \"3e1b036a-a1f1-44ed-9823-b49199c5b4de\") " pod="openstack/keystone-bootstrap-sz8zj" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.041165 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbsbx\" (UniqueName: \"kubernetes.io/projected/2dabdbea-d929-4b73-85a4-b6a882967131-kube-api-access-pbsbx\") pod \"dnsmasq-dns-6546db6db7-w7zk2\" (UID: \"2dabdbea-d929-4b73-85a4-b6a882967131\") " pod="openstack/dnsmasq-dns-6546db6db7-w7zk2" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.041226 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2dabdbea-d929-4b73-85a4-b6a882967131-ovsdbserver-sb\") pod \"dnsmasq-dns-6546db6db7-w7zk2\" (UID: \"2dabdbea-d929-4b73-85a4-b6a882967131\") " pod="openstack/dnsmasq-dns-6546db6db7-w7zk2" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.041293 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e1b036a-a1f1-44ed-9823-b49199c5b4de-config-data\") pod \"keystone-bootstrap-sz8zj\" (UID: \"3e1b036a-a1f1-44ed-9823-b49199c5b4de\") " pod="openstack/keystone-bootstrap-sz8zj" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.041311 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3e1b036a-a1f1-44ed-9823-b49199c5b4de-fernet-keys\") pod \"keystone-bootstrap-sz8zj\" (UID: \"3e1b036a-a1f1-44ed-9823-b49199c5b4de\") " pod="openstack/keystone-bootstrap-sz8zj" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.041344 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2dabdbea-d929-4b73-85a4-b6a882967131-dns-svc\") pod \"dnsmasq-dns-6546db6db7-w7zk2\" (UID: \"2dabdbea-d929-4b73-85a4-b6a882967131\") " pod="openstack/dnsmasq-dns-6546db6db7-w7zk2" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.041385 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e1b036a-a1f1-44ed-9823-b49199c5b4de-scripts\") pod \"keystone-bootstrap-sz8zj\" (UID: \"3e1b036a-a1f1-44ed-9823-b49199c5b4de\") " pod="openstack/keystone-bootstrap-sz8zj" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.041414 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kh29k\" (UniqueName: \"kubernetes.io/projected/3e1b036a-a1f1-44ed-9823-b49199c5b4de-kube-api-access-kh29k\") pod \"keystone-bootstrap-sz8zj\" (UID: \"3e1b036a-a1f1-44ed-9823-b49199c5b4de\") " pod="openstack/keystone-bootstrap-sz8zj" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.041430 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2dabdbea-d929-4b73-85a4-b6a882967131-ovsdbserver-nb\") pod \"dnsmasq-dns-6546db6db7-w7zk2\" (UID: \"2dabdbea-d929-4b73-85a4-b6a882967131\") " pod="openstack/dnsmasq-dns-6546db6db7-w7zk2" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.041454 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2dabdbea-d929-4b73-85a4-b6a882967131-config\") pod \"dnsmasq-dns-6546db6db7-w7zk2\" (UID: \"2dabdbea-d929-4b73-85a4-b6a882967131\") " pod="openstack/dnsmasq-dns-6546db6db7-w7zk2" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.041495 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e1b036a-a1f1-44ed-9823-b49199c5b4de-combined-ca-bundle\") pod \"keystone-bootstrap-sz8zj\" (UID: \"3e1b036a-a1f1-44ed-9823-b49199c5b4de\") " pod="openstack/keystone-bootstrap-sz8zj" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.041920 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-9grld"] Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.044631 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-9grld" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.050169 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2dabdbea-d929-4b73-85a4-b6a882967131-dns-svc\") pod \"dnsmasq-dns-6546db6db7-w7zk2\" (UID: \"2dabdbea-d929-4b73-85a4-b6a882967131\") " pod="openstack/dnsmasq-dns-6546db6db7-w7zk2" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.055263 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.055590 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.057184 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2dabdbea-d929-4b73-85a4-b6a882967131-ovsdbserver-sb\") pod \"dnsmasq-dns-6546db6db7-w7zk2\" (UID: \"2dabdbea-d929-4b73-85a4-b6a882967131\") " pod="openstack/dnsmasq-dns-6546db6db7-w7zk2" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.059415 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2dabdbea-d929-4b73-85a4-b6a882967131-ovsdbserver-nb\") pod \"dnsmasq-dns-6546db6db7-w7zk2\" (UID: \"2dabdbea-d929-4b73-85a4-b6a882967131\") " pod="openstack/dnsmasq-dns-6546db6db7-w7zk2" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.059777 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-mqltq" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.060674 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e1b036a-a1f1-44ed-9823-b49199c5b4de-combined-ca-bundle\") pod \"keystone-bootstrap-sz8zj\" (UID: \"3e1b036a-a1f1-44ed-9823-b49199c5b4de\") " pod="openstack/keystone-bootstrap-sz8zj" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.060698 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2dabdbea-d929-4b73-85a4-b6a882967131-config\") pod \"dnsmasq-dns-6546db6db7-w7zk2\" (UID: \"2dabdbea-d929-4b73-85a4-b6a882967131\") " pod="openstack/dnsmasq-dns-6546db6db7-w7zk2" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.063681 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3e1b036a-a1f1-44ed-9823-b49199c5b4de-fernet-keys\") pod \"keystone-bootstrap-sz8zj\" (UID: \"3e1b036a-a1f1-44ed-9823-b49199c5b4de\") " pod="openstack/keystone-bootstrap-sz8zj" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.075209 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e1b036a-a1f1-44ed-9823-b49199c5b4de-scripts\") pod \"keystone-bootstrap-sz8zj\" (UID: \"3e1b036a-a1f1-44ed-9823-b49199c5b4de\") " pod="openstack/keystone-bootstrap-sz8zj" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.075827 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e1b036a-a1f1-44ed-9823-b49199c5b4de-config-data\") pod \"keystone-bootstrap-sz8zj\" (UID: \"3e1b036a-a1f1-44ed-9823-b49199c5b4de\") " pod="openstack/keystone-bootstrap-sz8zj" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.082240 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3e1b036a-a1f1-44ed-9823-b49199c5b4de-credential-keys\") pod \"keystone-bootstrap-sz8zj\" (UID: \"3e1b036a-a1f1-44ed-9823-b49199c5b4de\") " pod="openstack/keystone-bootstrap-sz8zj" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.086911 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-9grld"] Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.093936 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kh29k\" (UniqueName: \"kubernetes.io/projected/3e1b036a-a1f1-44ed-9823-b49199c5b4de-kube-api-access-kh29k\") pod \"keystone-bootstrap-sz8zj\" (UID: \"3e1b036a-a1f1-44ed-9823-b49199c5b4de\") " pod="openstack/keystone-bootstrap-sz8zj" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.109750 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbsbx\" (UniqueName: \"kubernetes.io/projected/2dabdbea-d929-4b73-85a4-b6a882967131-kube-api-access-pbsbx\") pod \"dnsmasq-dns-6546db6db7-w7zk2\" (UID: \"2dabdbea-d929-4b73-85a4-b6a882967131\") " pod="openstack/dnsmasq-dns-6546db6db7-w7zk2" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.128885 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.131055 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.139517 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.139918 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.145525 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/bcf0c49a-5656-4399-837c-152c9af16309-config\") pod \"neutron-db-sync-9grld\" (UID: \"bcf0c49a-5656-4399-837c-152c9af16309\") " pod="openstack/neutron-db-sync-9grld" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.149675 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/515c81c7-10e3-4724-aafd-42431946c2c2-db-sync-config-data\") pod \"cinder-db-sync-btd8j\" (UID: \"515c81c7-10e3-4724-aafd-42431946c2c2\") " pod="openstack/cinder-db-sync-btd8j" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.150233 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2n72n\" (UniqueName: \"kubernetes.io/projected/bcf0c49a-5656-4399-837c-152c9af16309-kube-api-access-2n72n\") pod \"neutron-db-sync-9grld\" (UID: \"bcf0c49a-5656-4399-837c-152c9af16309\") " pod="openstack/neutron-db-sync-9grld" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.150397 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/515c81c7-10e3-4724-aafd-42431946c2c2-etc-machine-id\") pod \"cinder-db-sync-btd8j\" (UID: \"515c81c7-10e3-4724-aafd-42431946c2c2\") " pod="openstack/cinder-db-sync-btd8j" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.150445 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shprr\" (UniqueName: \"kubernetes.io/projected/515c81c7-10e3-4724-aafd-42431946c2c2-kube-api-access-shprr\") pod \"cinder-db-sync-btd8j\" (UID: \"515c81c7-10e3-4724-aafd-42431946c2c2\") " pod="openstack/cinder-db-sync-btd8j" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.150487 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcf0c49a-5656-4399-837c-152c9af16309-combined-ca-bundle\") pod \"neutron-db-sync-9grld\" (UID: \"bcf0c49a-5656-4399-837c-152c9af16309\") " pod="openstack/neutron-db-sync-9grld" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.150540 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/515c81c7-10e3-4724-aafd-42431946c2c2-combined-ca-bundle\") pod \"cinder-db-sync-btd8j\" (UID: \"515c81c7-10e3-4724-aafd-42431946c2c2\") " pod="openstack/cinder-db-sync-btd8j" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.150653 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/515c81c7-10e3-4724-aafd-42431946c2c2-scripts\") pod \"cinder-db-sync-btd8j\" (UID: \"515c81c7-10e3-4724-aafd-42431946c2c2\") " pod="openstack/cinder-db-sync-btd8j" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.150688 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/515c81c7-10e3-4724-aafd-42431946c2c2-config-data\") pod \"cinder-db-sync-btd8j\" (UID: \"515c81c7-10e3-4724-aafd-42431946c2c2\") " pod="openstack/cinder-db-sync-btd8j" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.179474 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.233025 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-gcfzn"] Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.234318 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-gcfzn" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.237429 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.237702 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-dhn66" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.251304 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-sz8zj" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.254956 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2n72n\" (UniqueName: \"kubernetes.io/projected/bcf0c49a-5656-4399-837c-152c9af16309-kube-api-access-2n72n\") pod \"neutron-db-sync-9grld\" (UID: \"bcf0c49a-5656-4399-837c-152c9af16309\") " pod="openstack/neutron-db-sync-9grld" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.255038 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/515c81c7-10e3-4724-aafd-42431946c2c2-etc-machine-id\") pod \"cinder-db-sync-btd8j\" (UID: \"515c81c7-10e3-4724-aafd-42431946c2c2\") " pod="openstack/cinder-db-sync-btd8j" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.255072 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/82c42edf-7b25-411f-b00f-95ed3e74bede-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"82c42edf-7b25-411f-b00f-95ed3e74bede\") " pod="openstack/ceilometer-0" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.255090 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/82c42edf-7b25-411f-b00f-95ed3e74bede-run-httpd\") pod \"ceilometer-0\" (UID: \"82c42edf-7b25-411f-b00f-95ed3e74bede\") " pod="openstack/ceilometer-0" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.255115 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shprr\" (UniqueName: \"kubernetes.io/projected/515c81c7-10e3-4724-aafd-42431946c2c2-kube-api-access-shprr\") pod \"cinder-db-sync-btd8j\" (UID: \"515c81c7-10e3-4724-aafd-42431946c2c2\") " pod="openstack/cinder-db-sync-btd8j" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.255141 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcf0c49a-5656-4399-837c-152c9af16309-combined-ca-bundle\") pod \"neutron-db-sync-9grld\" (UID: \"bcf0c49a-5656-4399-837c-152c9af16309\") " pod="openstack/neutron-db-sync-9grld" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.255171 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/515c81c7-10e3-4724-aafd-42431946c2c2-combined-ca-bundle\") pod \"cinder-db-sync-btd8j\" (UID: \"515c81c7-10e3-4724-aafd-42431946c2c2\") " pod="openstack/cinder-db-sync-btd8j" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.255190 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82c42edf-7b25-411f-b00f-95ed3e74bede-scripts\") pod \"ceilometer-0\" (UID: \"82c42edf-7b25-411f-b00f-95ed3e74bede\") " pod="openstack/ceilometer-0" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.255230 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/82c42edf-7b25-411f-b00f-95ed3e74bede-log-httpd\") pod \"ceilometer-0\" (UID: \"82c42edf-7b25-411f-b00f-95ed3e74bede\") " pod="openstack/ceilometer-0" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.255262 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/515c81c7-10e3-4724-aafd-42431946c2c2-scripts\") pod \"cinder-db-sync-btd8j\" (UID: \"515c81c7-10e3-4724-aafd-42431946c2c2\") " pod="openstack/cinder-db-sync-btd8j" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.255286 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/515c81c7-10e3-4724-aafd-42431946c2c2-config-data\") pod \"cinder-db-sync-btd8j\" (UID: \"515c81c7-10e3-4724-aafd-42431946c2c2\") " pod="openstack/cinder-db-sync-btd8j" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.255309 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82c42edf-7b25-411f-b00f-95ed3e74bede-config-data\") pod \"ceilometer-0\" (UID: \"82c42edf-7b25-411f-b00f-95ed3e74bede\") " pod="openstack/ceilometer-0" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.255339 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82c42edf-7b25-411f-b00f-95ed3e74bede-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"82c42edf-7b25-411f-b00f-95ed3e74bede\") " pod="openstack/ceilometer-0" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.255362 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/bcf0c49a-5656-4399-837c-152c9af16309-config\") pod \"neutron-db-sync-9grld\" (UID: \"bcf0c49a-5656-4399-837c-152c9af16309\") " pod="openstack/neutron-db-sync-9grld" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.255395 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zld5m\" (UniqueName: \"kubernetes.io/projected/82c42edf-7b25-411f-b00f-95ed3e74bede-kube-api-access-zld5m\") pod \"ceilometer-0\" (UID: \"82c42edf-7b25-411f-b00f-95ed3e74bede\") " pod="openstack/ceilometer-0" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.255415 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/515c81c7-10e3-4724-aafd-42431946c2c2-db-sync-config-data\") pod \"cinder-db-sync-btd8j\" (UID: \"515c81c7-10e3-4724-aafd-42431946c2c2\") " pod="openstack/cinder-db-sync-btd8j" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.255465 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/515c81c7-10e3-4724-aafd-42431946c2c2-etc-machine-id\") pod \"cinder-db-sync-btd8j\" (UID: \"515c81c7-10e3-4724-aafd-42431946c2c2\") " pod="openstack/cinder-db-sync-btd8j" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.266626 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/515c81c7-10e3-4724-aafd-42431946c2c2-scripts\") pod \"cinder-db-sync-btd8j\" (UID: \"515c81c7-10e3-4724-aafd-42431946c2c2\") " pod="openstack/cinder-db-sync-btd8j" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.266782 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcf0c49a-5656-4399-837c-152c9af16309-combined-ca-bundle\") pod \"neutron-db-sync-9grld\" (UID: \"bcf0c49a-5656-4399-837c-152c9af16309\") " pod="openstack/neutron-db-sync-9grld" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.275304 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/515c81c7-10e3-4724-aafd-42431946c2c2-db-sync-config-data\") pod \"cinder-db-sync-btd8j\" (UID: \"515c81c7-10e3-4724-aafd-42431946c2c2\") " pod="openstack/cinder-db-sync-btd8j" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.276836 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/bcf0c49a-5656-4399-837c-152c9af16309-config\") pod \"neutron-db-sync-9grld\" (UID: \"bcf0c49a-5656-4399-837c-152c9af16309\") " pod="openstack/neutron-db-sync-9grld" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.277324 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6546db6db7-w7zk2" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.279273 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-gcfzn"] Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.280548 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/515c81c7-10e3-4724-aafd-42431946c2c2-config-data\") pod \"cinder-db-sync-btd8j\" (UID: \"515c81c7-10e3-4724-aafd-42431946c2c2\") " pod="openstack/cinder-db-sync-btd8j" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.291294 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shprr\" (UniqueName: \"kubernetes.io/projected/515c81c7-10e3-4724-aafd-42431946c2c2-kube-api-access-shprr\") pod \"cinder-db-sync-btd8j\" (UID: \"515c81c7-10e3-4724-aafd-42431946c2c2\") " pod="openstack/cinder-db-sync-btd8j" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.291764 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2n72n\" (UniqueName: \"kubernetes.io/projected/bcf0c49a-5656-4399-837c-152c9af16309-kube-api-access-2n72n\") pod \"neutron-db-sync-9grld\" (UID: \"bcf0c49a-5656-4399-837c-152c9af16309\") " pod="openstack/neutron-db-sync-9grld" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.292293 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/515c81c7-10e3-4724-aafd-42431946c2c2-combined-ca-bundle\") pod \"cinder-db-sync-btd8j\" (UID: \"515c81c7-10e3-4724-aafd-42431946c2c2\") " pod="openstack/cinder-db-sync-btd8j" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.295881 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-6cz7h"] Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.297266 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-6cz7h" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.299668 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.299962 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.300054 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-mkdff" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.312885 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-6cz7h"] Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.358890 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-btd8j" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.360578 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zld5m\" (UniqueName: \"kubernetes.io/projected/82c42edf-7b25-411f-b00f-95ed3e74bede-kube-api-access-zld5m\") pod \"ceilometer-0\" (UID: \"82c42edf-7b25-411f-b00f-95ed3e74bede\") " pod="openstack/ceilometer-0" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.360676 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/82c42edf-7b25-411f-b00f-95ed3e74bede-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"82c42edf-7b25-411f-b00f-95ed3e74bede\") " pod="openstack/ceilometer-0" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.360700 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/82c42edf-7b25-411f-b00f-95ed3e74bede-run-httpd\") pod \"ceilometer-0\" (UID: \"82c42edf-7b25-411f-b00f-95ed3e74bede\") " pod="openstack/ceilometer-0" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.360738 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82c42edf-7b25-411f-b00f-95ed3e74bede-scripts\") pod \"ceilometer-0\" (UID: \"82c42edf-7b25-411f-b00f-95ed3e74bede\") " pod="openstack/ceilometer-0" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.360784 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgq57\" (UniqueName: \"kubernetes.io/projected/8370f287-fa77-4ef5-b929-01a4cf8c598c-kube-api-access-mgq57\") pod \"barbican-db-sync-gcfzn\" (UID: \"8370f287-fa77-4ef5-b929-01a4cf8c598c\") " pod="openstack/barbican-db-sync-gcfzn" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.361026 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/82c42edf-7b25-411f-b00f-95ed3e74bede-log-httpd\") pod \"ceilometer-0\" (UID: \"82c42edf-7b25-411f-b00f-95ed3e74bede\") " pod="openstack/ceilometer-0" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.361070 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8370f287-fa77-4ef5-b929-01a4cf8c598c-db-sync-config-data\") pod \"barbican-db-sync-gcfzn\" (UID: \"8370f287-fa77-4ef5-b929-01a4cf8c598c\") " pod="openstack/barbican-db-sync-gcfzn" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.361116 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82c42edf-7b25-411f-b00f-95ed3e74bede-config-data\") pod \"ceilometer-0\" (UID: \"82c42edf-7b25-411f-b00f-95ed3e74bede\") " pod="openstack/ceilometer-0" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.361150 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82c42edf-7b25-411f-b00f-95ed3e74bede-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"82c42edf-7b25-411f-b00f-95ed3e74bede\") " pod="openstack/ceilometer-0" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.361192 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8370f287-fa77-4ef5-b929-01a4cf8c598c-combined-ca-bundle\") pod \"barbican-db-sync-gcfzn\" (UID: \"8370f287-fa77-4ef5-b929-01a4cf8c598c\") " pod="openstack/barbican-db-sync-gcfzn" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.361744 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/82c42edf-7b25-411f-b00f-95ed3e74bede-run-httpd\") pod \"ceilometer-0\" (UID: \"82c42edf-7b25-411f-b00f-95ed3e74bede\") " pod="openstack/ceilometer-0" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.362902 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/82c42edf-7b25-411f-b00f-95ed3e74bede-log-httpd\") pod \"ceilometer-0\" (UID: \"82c42edf-7b25-411f-b00f-95ed3e74bede\") " pod="openstack/ceilometer-0" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.375188 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6546db6db7-w7zk2"] Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.382595 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82c42edf-7b25-411f-b00f-95ed3e74bede-scripts\") pod \"ceilometer-0\" (UID: \"82c42edf-7b25-411f-b00f-95ed3e74bede\") " pod="openstack/ceilometer-0" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.393850 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7987f74bbc-m7qr2"] Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.393883 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/82c42edf-7b25-411f-b00f-95ed3e74bede-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"82c42edf-7b25-411f-b00f-95ed3e74bede\") " pod="openstack/ceilometer-0" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.397123 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82c42edf-7b25-411f-b00f-95ed3e74bede-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"82c42edf-7b25-411f-b00f-95ed3e74bede\") " pod="openstack/ceilometer-0" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.397805 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82c42edf-7b25-411f-b00f-95ed3e74bede-config-data\") pod \"ceilometer-0\" (UID: \"82c42edf-7b25-411f-b00f-95ed3e74bede\") " pod="openstack/ceilometer-0" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.400021 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zld5m\" (UniqueName: \"kubernetes.io/projected/82c42edf-7b25-411f-b00f-95ed3e74bede-kube-api-access-zld5m\") pod \"ceilometer-0\" (UID: \"82c42edf-7b25-411f-b00f-95ed3e74bede\") " pod="openstack/ceilometer-0" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.401413 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7987f74bbc-m7qr2" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.421942 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-9grld" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.428266 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7987f74bbc-m7qr2"] Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.463680 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57a0c30f-28ae-466a-bc63-59ad7b6afd51-config-data\") pod \"placement-db-sync-6cz7h\" (UID: \"57a0c30f-28ae-466a-bc63-59ad7b6afd51\") " pod="openstack/placement-db-sync-6cz7h" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.464738 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57a0c30f-28ae-466a-bc63-59ad7b6afd51-scripts\") pod \"placement-db-sync-6cz7h\" (UID: \"57a0c30f-28ae-466a-bc63-59ad7b6afd51\") " pod="openstack/placement-db-sync-6cz7h" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.464774 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjqqw\" (UniqueName: \"kubernetes.io/projected/57a0c30f-28ae-466a-bc63-59ad7b6afd51-kube-api-access-zjqqw\") pod \"placement-db-sync-6cz7h\" (UID: \"57a0c30f-28ae-466a-bc63-59ad7b6afd51\") " pod="openstack/placement-db-sync-6cz7h" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.464845 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgq57\" (UniqueName: \"kubernetes.io/projected/8370f287-fa77-4ef5-b929-01a4cf8c598c-kube-api-access-mgq57\") pod \"barbican-db-sync-gcfzn\" (UID: \"8370f287-fa77-4ef5-b929-01a4cf8c598c\") " pod="openstack/barbican-db-sync-gcfzn" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.464867 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8370f287-fa77-4ef5-b929-01a4cf8c598c-db-sync-config-data\") pod \"barbican-db-sync-gcfzn\" (UID: \"8370f287-fa77-4ef5-b929-01a4cf8c598c\") " pod="openstack/barbican-db-sync-gcfzn" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.464916 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57a0c30f-28ae-466a-bc63-59ad7b6afd51-combined-ca-bundle\") pod \"placement-db-sync-6cz7h\" (UID: \"57a0c30f-28ae-466a-bc63-59ad7b6afd51\") " pod="openstack/placement-db-sync-6cz7h" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.464934 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/57a0c30f-28ae-466a-bc63-59ad7b6afd51-logs\") pod \"placement-db-sync-6cz7h\" (UID: \"57a0c30f-28ae-466a-bc63-59ad7b6afd51\") " pod="openstack/placement-db-sync-6cz7h" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.464959 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8370f287-fa77-4ef5-b929-01a4cf8c598c-combined-ca-bundle\") pod \"barbican-db-sync-gcfzn\" (UID: \"8370f287-fa77-4ef5-b929-01a4cf8c598c\") " pod="openstack/barbican-db-sync-gcfzn" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.467871 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.483619 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8370f287-fa77-4ef5-b929-01a4cf8c598c-combined-ca-bundle\") pod \"barbican-db-sync-gcfzn\" (UID: \"8370f287-fa77-4ef5-b929-01a4cf8c598c\") " pod="openstack/barbican-db-sync-gcfzn" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.486069 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8370f287-fa77-4ef5-b929-01a4cf8c598c-db-sync-config-data\") pod \"barbican-db-sync-gcfzn\" (UID: \"8370f287-fa77-4ef5-b929-01a4cf8c598c\") " pod="openstack/barbican-db-sync-gcfzn" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.489640 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgq57\" (UniqueName: \"kubernetes.io/projected/8370f287-fa77-4ef5-b929-01a4cf8c598c-kube-api-access-mgq57\") pod \"barbican-db-sync-gcfzn\" (UID: \"8370f287-fa77-4ef5-b929-01a4cf8c598c\") " pod="openstack/barbican-db-sync-gcfzn" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.567415 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6n5x\" (UniqueName: \"kubernetes.io/projected/b80b0f55-9641-465a-b40c-f90244e53218-kube-api-access-s6n5x\") pod \"dnsmasq-dns-7987f74bbc-m7qr2\" (UID: \"b80b0f55-9641-465a-b40c-f90244e53218\") " pod="openstack/dnsmasq-dns-7987f74bbc-m7qr2" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.567506 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57a0c30f-28ae-466a-bc63-59ad7b6afd51-combined-ca-bundle\") pod \"placement-db-sync-6cz7h\" (UID: \"57a0c30f-28ae-466a-bc63-59ad7b6afd51\") " pod="openstack/placement-db-sync-6cz7h" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.567538 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/57a0c30f-28ae-466a-bc63-59ad7b6afd51-logs\") pod \"placement-db-sync-6cz7h\" (UID: \"57a0c30f-28ae-466a-bc63-59ad7b6afd51\") " pod="openstack/placement-db-sync-6cz7h" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.567574 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b80b0f55-9641-465a-b40c-f90244e53218-ovsdbserver-sb\") pod \"dnsmasq-dns-7987f74bbc-m7qr2\" (UID: \"b80b0f55-9641-465a-b40c-f90244e53218\") " pod="openstack/dnsmasq-dns-7987f74bbc-m7qr2" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.567600 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b80b0f55-9641-465a-b40c-f90244e53218-ovsdbserver-nb\") pod \"dnsmasq-dns-7987f74bbc-m7qr2\" (UID: \"b80b0f55-9641-465a-b40c-f90244e53218\") " pod="openstack/dnsmasq-dns-7987f74bbc-m7qr2" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.567627 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b80b0f55-9641-465a-b40c-f90244e53218-dns-svc\") pod \"dnsmasq-dns-7987f74bbc-m7qr2\" (UID: \"b80b0f55-9641-465a-b40c-f90244e53218\") " pod="openstack/dnsmasq-dns-7987f74bbc-m7qr2" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.567652 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b80b0f55-9641-465a-b40c-f90244e53218-config\") pod \"dnsmasq-dns-7987f74bbc-m7qr2\" (UID: \"b80b0f55-9641-465a-b40c-f90244e53218\") " pod="openstack/dnsmasq-dns-7987f74bbc-m7qr2" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.567685 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57a0c30f-28ae-466a-bc63-59ad7b6afd51-config-data\") pod \"placement-db-sync-6cz7h\" (UID: \"57a0c30f-28ae-466a-bc63-59ad7b6afd51\") " pod="openstack/placement-db-sync-6cz7h" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.567728 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57a0c30f-28ae-466a-bc63-59ad7b6afd51-scripts\") pod \"placement-db-sync-6cz7h\" (UID: \"57a0c30f-28ae-466a-bc63-59ad7b6afd51\") " pod="openstack/placement-db-sync-6cz7h" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.567748 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjqqw\" (UniqueName: \"kubernetes.io/projected/57a0c30f-28ae-466a-bc63-59ad7b6afd51-kube-api-access-zjqqw\") pod \"placement-db-sync-6cz7h\" (UID: \"57a0c30f-28ae-466a-bc63-59ad7b6afd51\") " pod="openstack/placement-db-sync-6cz7h" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.574893 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57a0c30f-28ae-466a-bc63-59ad7b6afd51-config-data\") pod \"placement-db-sync-6cz7h\" (UID: \"57a0c30f-28ae-466a-bc63-59ad7b6afd51\") " pod="openstack/placement-db-sync-6cz7h" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.578580 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57a0c30f-28ae-466a-bc63-59ad7b6afd51-scripts\") pod \"placement-db-sync-6cz7h\" (UID: \"57a0c30f-28ae-466a-bc63-59ad7b6afd51\") " pod="openstack/placement-db-sync-6cz7h" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.579451 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/57a0c30f-28ae-466a-bc63-59ad7b6afd51-logs\") pod \"placement-db-sync-6cz7h\" (UID: \"57a0c30f-28ae-466a-bc63-59ad7b6afd51\") " pod="openstack/placement-db-sync-6cz7h" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.579760 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57a0c30f-28ae-466a-bc63-59ad7b6afd51-combined-ca-bundle\") pod \"placement-db-sync-6cz7h\" (UID: \"57a0c30f-28ae-466a-bc63-59ad7b6afd51\") " pod="openstack/placement-db-sync-6cz7h" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.581570 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54f9b7b8d9-8sxd6" event={"ID":"64328f76-389c-43d8-9f70-e6888983268e","Type":"ContainerStarted","Data":"330a6e364efd144e951da6191f9e5fe358937120b286aa57e8f9021641d04c57"} Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.592529 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjqqw\" (UniqueName: \"kubernetes.io/projected/57a0c30f-28ae-466a-bc63-59ad7b6afd51-kube-api-access-zjqqw\") pod \"placement-db-sync-6cz7h\" (UID: \"57a0c30f-28ae-466a-bc63-59ad7b6afd51\") " pod="openstack/placement-db-sync-6cz7h" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.669250 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6n5x\" (UniqueName: \"kubernetes.io/projected/b80b0f55-9641-465a-b40c-f90244e53218-kube-api-access-s6n5x\") pod \"dnsmasq-dns-7987f74bbc-m7qr2\" (UID: \"b80b0f55-9641-465a-b40c-f90244e53218\") " pod="openstack/dnsmasq-dns-7987f74bbc-m7qr2" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.670491 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b80b0f55-9641-465a-b40c-f90244e53218-ovsdbserver-sb\") pod \"dnsmasq-dns-7987f74bbc-m7qr2\" (UID: \"b80b0f55-9641-465a-b40c-f90244e53218\") " pod="openstack/dnsmasq-dns-7987f74bbc-m7qr2" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.670722 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b80b0f55-9641-465a-b40c-f90244e53218-ovsdbserver-nb\") pod \"dnsmasq-dns-7987f74bbc-m7qr2\" (UID: \"b80b0f55-9641-465a-b40c-f90244e53218\") " pod="openstack/dnsmasq-dns-7987f74bbc-m7qr2" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.670974 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b80b0f55-9641-465a-b40c-f90244e53218-dns-svc\") pod \"dnsmasq-dns-7987f74bbc-m7qr2\" (UID: \"b80b0f55-9641-465a-b40c-f90244e53218\") " pod="openstack/dnsmasq-dns-7987f74bbc-m7qr2" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.671035 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b80b0f55-9641-465a-b40c-f90244e53218-config\") pod \"dnsmasq-dns-7987f74bbc-m7qr2\" (UID: \"b80b0f55-9641-465a-b40c-f90244e53218\") " pod="openstack/dnsmasq-dns-7987f74bbc-m7qr2" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.671902 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b80b0f55-9641-465a-b40c-f90244e53218-ovsdbserver-sb\") pod \"dnsmasq-dns-7987f74bbc-m7qr2\" (UID: \"b80b0f55-9641-465a-b40c-f90244e53218\") " pod="openstack/dnsmasq-dns-7987f74bbc-m7qr2" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.672215 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b80b0f55-9641-465a-b40c-f90244e53218-ovsdbserver-nb\") pod \"dnsmasq-dns-7987f74bbc-m7qr2\" (UID: \"b80b0f55-9641-465a-b40c-f90244e53218\") " pod="openstack/dnsmasq-dns-7987f74bbc-m7qr2" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.672599 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b80b0f55-9641-465a-b40c-f90244e53218-dns-svc\") pod \"dnsmasq-dns-7987f74bbc-m7qr2\" (UID: \"b80b0f55-9641-465a-b40c-f90244e53218\") " pod="openstack/dnsmasq-dns-7987f74bbc-m7qr2" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.673193 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b80b0f55-9641-465a-b40c-f90244e53218-config\") pod \"dnsmasq-dns-7987f74bbc-m7qr2\" (UID: \"b80b0f55-9641-465a-b40c-f90244e53218\") " pod="openstack/dnsmasq-dns-7987f74bbc-m7qr2" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.698002 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-gcfzn" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.704105 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6n5x\" (UniqueName: \"kubernetes.io/projected/b80b0f55-9641-465a-b40c-f90244e53218-kube-api-access-s6n5x\") pod \"dnsmasq-dns-7987f74bbc-m7qr2\" (UID: \"b80b0f55-9641-465a-b40c-f90244e53218\") " pod="openstack/dnsmasq-dns-7987f74bbc-m7qr2" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.715130 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-6cz7h" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.725895 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7987f74bbc-m7qr2" Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.911714 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-sz8zj"] Nov 29 04:54:57 crc kubenswrapper[4799]: W1129 04:54:57.920778 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3e1b036a_a1f1_44ed_9823_b49199c5b4de.slice/crio-36fafcbe0ace23e99b5007d42af4eebf4548a5e8f190ea49d398d56790f2d4ba WatchSource:0}: Error finding container 36fafcbe0ace23e99b5007d42af4eebf4548a5e8f190ea49d398d56790f2d4ba: Status 404 returned error can't find the container with id 36fafcbe0ace23e99b5007d42af4eebf4548a5e8f190ea49d398d56790f2d4ba Nov 29 04:54:57 crc kubenswrapper[4799]: I1129 04:54:57.983625 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6546db6db7-w7zk2"] Nov 29 04:54:57 crc kubenswrapper[4799]: W1129 04:54:57.987398 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2dabdbea_d929_4b73_85a4_b6a882967131.slice/crio-a6c5c7b52c9f0084f433ff2cfa4dc80d395c740e7c38e2d872788f81570f1f45 WatchSource:0}: Error finding container a6c5c7b52c9f0084f433ff2cfa4dc80d395c740e7c38e2d872788f81570f1f45: Status 404 returned error can't find the container with id a6c5c7b52c9f0084f433ff2cfa4dc80d395c740e7c38e2d872788f81570f1f45 Nov 29 04:54:58 crc kubenswrapper[4799]: I1129 04:54:58.077289 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-9grld"] Nov 29 04:54:58 crc kubenswrapper[4799]: I1129 04:54:58.100247 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-btd8j"] Nov 29 04:54:58 crc kubenswrapper[4799]: W1129 04:54:58.107942 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod515c81c7_10e3_4724_aafd_42431946c2c2.slice/crio-8d5dc91c375bc0e9b87529b6d8e195a562d55d9e08494ad48c856841aeeb21df WatchSource:0}: Error finding container 8d5dc91c375bc0e9b87529b6d8e195a562d55d9e08494ad48c856841aeeb21df: Status 404 returned error can't find the container with id 8d5dc91c375bc0e9b87529b6d8e195a562d55d9e08494ad48c856841aeeb21df Nov 29 04:54:58 crc kubenswrapper[4799]: I1129 04:54:58.110395 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 29 04:54:58 crc kubenswrapper[4799]: I1129 04:54:58.344649 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-gcfzn"] Nov 29 04:54:58 crc kubenswrapper[4799]: I1129 04:54:58.360378 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7987f74bbc-m7qr2"] Nov 29 04:54:58 crc kubenswrapper[4799]: W1129 04:54:58.373136 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8370f287_fa77_4ef5_b929_01a4cf8c598c.slice/crio-6b706758924ab4618bd3d2d352c81159df1871030ddea09a02a3f8a7e722ffd4 WatchSource:0}: Error finding container 6b706758924ab4618bd3d2d352c81159df1871030ddea09a02a3f8a7e722ffd4: Status 404 returned error can't find the container with id 6b706758924ab4618bd3d2d352c81159df1871030ddea09a02a3f8a7e722ffd4 Nov 29 04:54:58 crc kubenswrapper[4799]: I1129 04:54:58.374463 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-6cz7h"] Nov 29 04:54:58 crc kubenswrapper[4799]: W1129 04:54:58.389165 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb80b0f55_9641_465a_b40c_f90244e53218.slice/crio-53a604240972478ff9052bfa352a9d253572828d489daca54c06e91ebb6b4e85 WatchSource:0}: Error finding container 53a604240972478ff9052bfa352a9d253572828d489daca54c06e91ebb6b4e85: Status 404 returned error can't find the container with id 53a604240972478ff9052bfa352a9d253572828d489daca54c06e91ebb6b4e85 Nov 29 04:54:58 crc kubenswrapper[4799]: W1129 04:54:58.395288 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod57a0c30f_28ae_466a_bc63_59ad7b6afd51.slice/crio-1474247c5094f84feb0e2529060061812c7fe23b552ef19e6ff9efa77463600f WatchSource:0}: Error finding container 1474247c5094f84feb0e2529060061812c7fe23b552ef19e6ff9efa77463600f: Status 404 returned error can't find the container with id 1474247c5094f84feb0e2529060061812c7fe23b552ef19e6ff9efa77463600f Nov 29 04:54:58 crc kubenswrapper[4799]: I1129 04:54:58.591140 4799 generic.go:334] "Generic (PLEG): container finished" podID="2dabdbea-d929-4b73-85a4-b6a882967131" containerID="ee420b24827233c65c86eced02e2602848b88fc0f6823a8a21f1a739d18771f7" exitCode=0 Nov 29 04:54:58 crc kubenswrapper[4799]: I1129 04:54:58.591247 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6546db6db7-w7zk2" event={"ID":"2dabdbea-d929-4b73-85a4-b6a882967131","Type":"ContainerDied","Data":"ee420b24827233c65c86eced02e2602848b88fc0f6823a8a21f1a739d18771f7"} Nov 29 04:54:58 crc kubenswrapper[4799]: I1129 04:54:58.591307 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6546db6db7-w7zk2" event={"ID":"2dabdbea-d929-4b73-85a4-b6a882967131","Type":"ContainerStarted","Data":"a6c5c7b52c9f0084f433ff2cfa4dc80d395c740e7c38e2d872788f81570f1f45"} Nov 29 04:54:58 crc kubenswrapper[4799]: I1129 04:54:58.594283 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-btd8j" event={"ID":"515c81c7-10e3-4724-aafd-42431946c2c2","Type":"ContainerStarted","Data":"8d5dc91c375bc0e9b87529b6d8e195a562d55d9e08494ad48c856841aeeb21df"} Nov 29 04:54:58 crc kubenswrapper[4799]: I1129 04:54:58.597017 4799 generic.go:334] "Generic (PLEG): container finished" podID="64328f76-389c-43d8-9f70-e6888983268e" containerID="6d425ecfb1a047cbc72185dce1b126488dec749e7690524319b4010f1116b35f" exitCode=0 Nov 29 04:54:58 crc kubenswrapper[4799]: I1129 04:54:58.597273 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54f9b7b8d9-8sxd6" event={"ID":"64328f76-389c-43d8-9f70-e6888983268e","Type":"ContainerDied","Data":"6d425ecfb1a047cbc72185dce1b126488dec749e7690524319b4010f1116b35f"} Nov 29 04:54:58 crc kubenswrapper[4799]: I1129 04:54:58.599367 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-sz8zj" event={"ID":"3e1b036a-a1f1-44ed-9823-b49199c5b4de","Type":"ContainerStarted","Data":"9bb14bd0f51cef201cbf42d18d47d64600f1916ddc5961f149e18335ea33cc59"} Nov 29 04:54:58 crc kubenswrapper[4799]: I1129 04:54:58.599405 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-sz8zj" event={"ID":"3e1b036a-a1f1-44ed-9823-b49199c5b4de","Type":"ContainerStarted","Data":"36fafcbe0ace23e99b5007d42af4eebf4548a5e8f190ea49d398d56790f2d4ba"} Nov 29 04:54:58 crc kubenswrapper[4799]: I1129 04:54:58.602089 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"82c42edf-7b25-411f-b00f-95ed3e74bede","Type":"ContainerStarted","Data":"107ffbdcfced5b102126423c5623d97eae4be041870337b29984647a277720a5"} Nov 29 04:54:58 crc kubenswrapper[4799]: I1129 04:54:58.604355 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-9grld" event={"ID":"bcf0c49a-5656-4399-837c-152c9af16309","Type":"ContainerStarted","Data":"cdcd2d082a36adb9d24d954ba9b89729c3c908a3e760003812816af54bcbb49c"} Nov 29 04:54:58 crc kubenswrapper[4799]: I1129 04:54:58.604390 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-9grld" event={"ID":"bcf0c49a-5656-4399-837c-152c9af16309","Type":"ContainerStarted","Data":"9403fb2aec9ba9689eb737b70975060b393fe7471cb57ab3f502c4abbb198fc2"} Nov 29 04:54:58 crc kubenswrapper[4799]: I1129 04:54:58.612835 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-6cz7h" event={"ID":"57a0c30f-28ae-466a-bc63-59ad7b6afd51","Type":"ContainerStarted","Data":"1474247c5094f84feb0e2529060061812c7fe23b552ef19e6ff9efa77463600f"} Nov 29 04:54:58 crc kubenswrapper[4799]: I1129 04:54:58.628858 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-gcfzn" event={"ID":"8370f287-fa77-4ef5-b929-01a4cf8c598c","Type":"ContainerStarted","Data":"6b706758924ab4618bd3d2d352c81159df1871030ddea09a02a3f8a7e722ffd4"} Nov 29 04:54:58 crc kubenswrapper[4799]: I1129 04:54:58.633166 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7987f74bbc-m7qr2" event={"ID":"b80b0f55-9641-465a-b40c-f90244e53218","Type":"ContainerStarted","Data":"53a604240972478ff9052bfa352a9d253572828d489daca54c06e91ebb6b4e85"} Nov 29 04:54:58 crc kubenswrapper[4799]: I1129 04:54:58.695570 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-9grld" podStartSLOduration=2.695542187 podStartE2EDuration="2.695542187s" podCreationTimestamp="2025-11-29 04:54:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:54:58.690324519 +0000 UTC m=+974.333254919" watchObservedRunningTime="2025-11-29 04:54:58.695542187 +0000 UTC m=+974.338472597" Nov 29 04:54:58 crc kubenswrapper[4799]: I1129 04:54:58.703990 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-sz8zj" podStartSLOduration=2.703972663 podStartE2EDuration="2.703972663s" podCreationTimestamp="2025-11-29 04:54:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:54:58.669492972 +0000 UTC m=+974.312423372" watchObservedRunningTime="2025-11-29 04:54:58.703972663 +0000 UTC m=+974.346903063" Nov 29 04:54:58 crc kubenswrapper[4799]: I1129 04:54:58.931039 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54f9b7b8d9-8sxd6" Nov 29 04:54:58 crc kubenswrapper[4799]: I1129 04:54:58.997904 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6546db6db7-w7zk2" Nov 29 04:54:59 crc kubenswrapper[4799]: I1129 04:54:59.115606 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/64328f76-389c-43d8-9f70-e6888983268e-dns-svc\") pod \"64328f76-389c-43d8-9f70-e6888983268e\" (UID: \"64328f76-389c-43d8-9f70-e6888983268e\") " Nov 29 04:54:59 crc kubenswrapper[4799]: I1129 04:54:59.115699 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/64328f76-389c-43d8-9f70-e6888983268e-ovsdbserver-nb\") pod \"64328f76-389c-43d8-9f70-e6888983268e\" (UID: \"64328f76-389c-43d8-9f70-e6888983268e\") " Nov 29 04:54:59 crc kubenswrapper[4799]: I1129 04:54:59.115735 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2dabdbea-d929-4b73-85a4-b6a882967131-config\") pod \"2dabdbea-d929-4b73-85a4-b6a882967131\" (UID: \"2dabdbea-d929-4b73-85a4-b6a882967131\") " Nov 29 04:54:59 crc kubenswrapper[4799]: I1129 04:54:59.115766 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2dabdbea-d929-4b73-85a4-b6a882967131-dns-svc\") pod \"2dabdbea-d929-4b73-85a4-b6a882967131\" (UID: \"2dabdbea-d929-4b73-85a4-b6a882967131\") " Nov 29 04:54:59 crc kubenswrapper[4799]: I1129 04:54:59.115827 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2dabdbea-d929-4b73-85a4-b6a882967131-ovsdbserver-nb\") pod \"2dabdbea-d929-4b73-85a4-b6a882967131\" (UID: \"2dabdbea-d929-4b73-85a4-b6a882967131\") " Nov 29 04:54:59 crc kubenswrapper[4799]: I1129 04:54:59.115984 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pbsbx\" (UniqueName: \"kubernetes.io/projected/2dabdbea-d929-4b73-85a4-b6a882967131-kube-api-access-pbsbx\") pod \"2dabdbea-d929-4b73-85a4-b6a882967131\" (UID: \"2dabdbea-d929-4b73-85a4-b6a882967131\") " Nov 29 04:54:59 crc kubenswrapper[4799]: I1129 04:54:59.116011 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2dabdbea-d929-4b73-85a4-b6a882967131-ovsdbserver-sb\") pod \"2dabdbea-d929-4b73-85a4-b6a882967131\" (UID: \"2dabdbea-d929-4b73-85a4-b6a882967131\") " Nov 29 04:54:59 crc kubenswrapper[4799]: I1129 04:54:59.116135 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hbj6n\" (UniqueName: \"kubernetes.io/projected/64328f76-389c-43d8-9f70-e6888983268e-kube-api-access-hbj6n\") pod \"64328f76-389c-43d8-9f70-e6888983268e\" (UID: \"64328f76-389c-43d8-9f70-e6888983268e\") " Nov 29 04:54:59 crc kubenswrapper[4799]: I1129 04:54:59.116176 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/64328f76-389c-43d8-9f70-e6888983268e-ovsdbserver-sb\") pod \"64328f76-389c-43d8-9f70-e6888983268e\" (UID: \"64328f76-389c-43d8-9f70-e6888983268e\") " Nov 29 04:54:59 crc kubenswrapper[4799]: I1129 04:54:59.116194 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64328f76-389c-43d8-9f70-e6888983268e-config\") pod \"64328f76-389c-43d8-9f70-e6888983268e\" (UID: \"64328f76-389c-43d8-9f70-e6888983268e\") " Nov 29 04:54:59 crc kubenswrapper[4799]: I1129 04:54:59.125870 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2dabdbea-d929-4b73-85a4-b6a882967131-kube-api-access-pbsbx" (OuterVolumeSpecName: "kube-api-access-pbsbx") pod "2dabdbea-d929-4b73-85a4-b6a882967131" (UID: "2dabdbea-d929-4b73-85a4-b6a882967131"). InnerVolumeSpecName "kube-api-access-pbsbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:54:59 crc kubenswrapper[4799]: I1129 04:54:59.127984 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64328f76-389c-43d8-9f70-e6888983268e-kube-api-access-hbj6n" (OuterVolumeSpecName: "kube-api-access-hbj6n") pod "64328f76-389c-43d8-9f70-e6888983268e" (UID: "64328f76-389c-43d8-9f70-e6888983268e"). InnerVolumeSpecName "kube-api-access-hbj6n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:54:59 crc kubenswrapper[4799]: I1129 04:54:59.147286 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64328f76-389c-43d8-9f70-e6888983268e-config" (OuterVolumeSpecName: "config") pod "64328f76-389c-43d8-9f70-e6888983268e" (UID: "64328f76-389c-43d8-9f70-e6888983268e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:54:59 crc kubenswrapper[4799]: I1129 04:54:59.154865 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64328f76-389c-43d8-9f70-e6888983268e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "64328f76-389c-43d8-9f70-e6888983268e" (UID: "64328f76-389c-43d8-9f70-e6888983268e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:54:59 crc kubenswrapper[4799]: I1129 04:54:59.155352 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2dabdbea-d929-4b73-85a4-b6a882967131-config" (OuterVolumeSpecName: "config") pod "2dabdbea-d929-4b73-85a4-b6a882967131" (UID: "2dabdbea-d929-4b73-85a4-b6a882967131"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:54:59 crc kubenswrapper[4799]: I1129 04:54:59.156830 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2dabdbea-d929-4b73-85a4-b6a882967131-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2dabdbea-d929-4b73-85a4-b6a882967131" (UID: "2dabdbea-d929-4b73-85a4-b6a882967131"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:54:59 crc kubenswrapper[4799]: I1129 04:54:59.159145 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2dabdbea-d929-4b73-85a4-b6a882967131-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2dabdbea-d929-4b73-85a4-b6a882967131" (UID: "2dabdbea-d929-4b73-85a4-b6a882967131"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:54:59 crc kubenswrapper[4799]: I1129 04:54:59.160550 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2dabdbea-d929-4b73-85a4-b6a882967131-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2dabdbea-d929-4b73-85a4-b6a882967131" (UID: "2dabdbea-d929-4b73-85a4-b6a882967131"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:54:59 crc kubenswrapper[4799]: I1129 04:54:59.162967 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64328f76-389c-43d8-9f70-e6888983268e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "64328f76-389c-43d8-9f70-e6888983268e" (UID: "64328f76-389c-43d8-9f70-e6888983268e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:54:59 crc kubenswrapper[4799]: I1129 04:54:59.165507 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64328f76-389c-43d8-9f70-e6888983268e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "64328f76-389c-43d8-9f70-e6888983268e" (UID: "64328f76-389c-43d8-9f70-e6888983268e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:54:59 crc kubenswrapper[4799]: I1129 04:54:59.219038 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pbsbx\" (UniqueName: \"kubernetes.io/projected/2dabdbea-d929-4b73-85a4-b6a882967131-kube-api-access-pbsbx\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:59 crc kubenswrapper[4799]: I1129 04:54:59.219098 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2dabdbea-d929-4b73-85a4-b6a882967131-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:59 crc kubenswrapper[4799]: I1129 04:54:59.219114 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hbj6n\" (UniqueName: \"kubernetes.io/projected/64328f76-389c-43d8-9f70-e6888983268e-kube-api-access-hbj6n\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:59 crc kubenswrapper[4799]: I1129 04:54:59.219128 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/64328f76-389c-43d8-9f70-e6888983268e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:59 crc kubenswrapper[4799]: I1129 04:54:59.219144 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64328f76-389c-43d8-9f70-e6888983268e-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:59 crc kubenswrapper[4799]: I1129 04:54:59.219160 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/64328f76-389c-43d8-9f70-e6888983268e-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:59 crc kubenswrapper[4799]: I1129 04:54:59.219174 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/64328f76-389c-43d8-9f70-e6888983268e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:59 crc kubenswrapper[4799]: I1129 04:54:59.219217 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2dabdbea-d929-4b73-85a4-b6a882967131-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:59 crc kubenswrapper[4799]: I1129 04:54:59.219231 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2dabdbea-d929-4b73-85a4-b6a882967131-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:59 crc kubenswrapper[4799]: I1129 04:54:59.219242 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2dabdbea-d929-4b73-85a4-b6a882967131-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 29 04:54:59 crc kubenswrapper[4799]: I1129 04:54:59.671974 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6546db6db7-w7zk2" event={"ID":"2dabdbea-d929-4b73-85a4-b6a882967131","Type":"ContainerDied","Data":"a6c5c7b52c9f0084f433ff2cfa4dc80d395c740e7c38e2d872788f81570f1f45"} Nov 29 04:54:59 crc kubenswrapper[4799]: I1129 04:54:59.672590 4799 scope.go:117] "RemoveContainer" containerID="ee420b24827233c65c86eced02e2602848b88fc0f6823a8a21f1a739d18771f7" Nov 29 04:54:59 crc kubenswrapper[4799]: I1129 04:54:59.672091 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6546db6db7-w7zk2" Nov 29 04:54:59 crc kubenswrapper[4799]: I1129 04:54:59.692297 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54f9b7b8d9-8sxd6" event={"ID":"64328f76-389c-43d8-9f70-e6888983268e","Type":"ContainerDied","Data":"330a6e364efd144e951da6191f9e5fe358937120b286aa57e8f9021641d04c57"} Nov 29 04:54:59 crc kubenswrapper[4799]: I1129 04:54:59.692460 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54f9b7b8d9-8sxd6" Nov 29 04:54:59 crc kubenswrapper[4799]: I1129 04:54:59.713162 4799 generic.go:334] "Generic (PLEG): container finished" podID="b80b0f55-9641-465a-b40c-f90244e53218" containerID="0d0bf418898e7dc12fb2eecab71f5d4ba5cc8ee5f55a9bbbaeef94e5766137ea" exitCode=0 Nov 29 04:54:59 crc kubenswrapper[4799]: I1129 04:54:59.714116 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7987f74bbc-m7qr2" event={"ID":"b80b0f55-9641-465a-b40c-f90244e53218","Type":"ContainerDied","Data":"0d0bf418898e7dc12fb2eecab71f5d4ba5cc8ee5f55a9bbbaeef94e5766137ea"} Nov 29 04:54:59 crc kubenswrapper[4799]: I1129 04:54:59.753540 4799 scope.go:117] "RemoveContainer" containerID="6d425ecfb1a047cbc72185dce1b126488dec749e7690524319b4010f1116b35f" Nov 29 04:54:59 crc kubenswrapper[4799]: I1129 04:54:59.803763 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6546db6db7-w7zk2"] Nov 29 04:54:59 crc kubenswrapper[4799]: I1129 04:54:59.828115 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6546db6db7-w7zk2"] Nov 29 04:54:59 crc kubenswrapper[4799]: I1129 04:54:59.887912 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54f9b7b8d9-8sxd6"] Nov 29 04:54:59 crc kubenswrapper[4799]: I1129 04:54:59.892761 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-54f9b7b8d9-8sxd6"] Nov 29 04:55:00 crc kubenswrapper[4799]: I1129 04:55:00.625966 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 29 04:55:00 crc kubenswrapper[4799]: I1129 04:55:00.677124 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2dabdbea-d929-4b73-85a4-b6a882967131" path="/var/lib/kubelet/pods/2dabdbea-d929-4b73-85a4-b6a882967131/volumes" Nov 29 04:55:00 crc kubenswrapper[4799]: I1129 04:55:00.677878 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64328f76-389c-43d8-9f70-e6888983268e" path="/var/lib/kubelet/pods/64328f76-389c-43d8-9f70-e6888983268e/volumes" Nov 29 04:55:00 crc kubenswrapper[4799]: I1129 04:55:00.733967 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7987f74bbc-m7qr2" event={"ID":"b80b0f55-9641-465a-b40c-f90244e53218","Type":"ContainerStarted","Data":"53fdab9dd5d6c0e0517ca438e2da7bb79dd9d0c61301ad55fbe44c10d61c1015"} Nov 29 04:55:00 crc kubenswrapper[4799]: I1129 04:55:00.734247 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7987f74bbc-m7qr2" Nov 29 04:55:00 crc kubenswrapper[4799]: I1129 04:55:00.769552 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7987f74bbc-m7qr2" podStartSLOduration=3.769528255 podStartE2EDuration="3.769528255s" podCreationTimestamp="2025-11-29 04:54:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:55:00.754822077 +0000 UTC m=+976.397752497" watchObservedRunningTime="2025-11-29 04:55:00.769528255 +0000 UTC m=+976.412458655" Nov 29 04:55:02 crc kubenswrapper[4799]: I1129 04:55:02.809352 4799 generic.go:334] "Generic (PLEG): container finished" podID="3e1b036a-a1f1-44ed-9823-b49199c5b4de" containerID="9bb14bd0f51cef201cbf42d18d47d64600f1916ddc5961f149e18335ea33cc59" exitCode=0 Nov 29 04:55:02 crc kubenswrapper[4799]: I1129 04:55:02.810124 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-sz8zj" event={"ID":"3e1b036a-a1f1-44ed-9823-b49199c5b4de","Type":"ContainerDied","Data":"9bb14bd0f51cef201cbf42d18d47d64600f1916ddc5961f149e18335ea33cc59"} Nov 29 04:55:07 crc kubenswrapper[4799]: I1129 04:55:07.498271 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 04:55:07 crc kubenswrapper[4799]: I1129 04:55:07.499149 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 04:55:07 crc kubenswrapper[4799]: I1129 04:55:07.499211 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-26t88" Nov 29 04:55:07 crc kubenswrapper[4799]: I1129 04:55:07.500090 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8a07429ac3c8e7bd45ed97e1ba2d6c50f73c78cd13fcb21a144bfb0b7bc995f6"} pod="openshift-machine-config-operator/machine-config-daemon-26t88" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 29 04:55:07 crc kubenswrapper[4799]: I1129 04:55:07.500155 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" containerID="cri-o://8a07429ac3c8e7bd45ed97e1ba2d6c50f73c78cd13fcb21a144bfb0b7bc995f6" gracePeriod=600 Nov 29 04:55:07 crc kubenswrapper[4799]: I1129 04:55:07.728094 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7987f74bbc-m7qr2" Nov 29 04:55:07 crc kubenswrapper[4799]: I1129 04:55:07.843566 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-dbv6d"] Nov 29 04:55:07 crc kubenswrapper[4799]: I1129 04:55:07.844500 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-dbv6d" podUID="82f8e690-e10d-4852-bc45-beba0e5b666e" containerName="dnsmasq-dns" containerID="cri-o://7d8eadeceea35e76b407cce3a32ef5e187792093f406f2a855c2100538b72f7b" gracePeriod=10 Nov 29 04:55:07 crc kubenswrapper[4799]: I1129 04:55:07.900414 4799 generic.go:334] "Generic (PLEG): container finished" podID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerID="8a07429ac3c8e7bd45ed97e1ba2d6c50f73c78cd13fcb21a144bfb0b7bc995f6" exitCode=0 Nov 29 04:55:07 crc kubenswrapper[4799]: I1129 04:55:07.900472 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" event={"ID":"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8","Type":"ContainerDied","Data":"8a07429ac3c8e7bd45ed97e1ba2d6c50f73c78cd13fcb21a144bfb0b7bc995f6"} Nov 29 04:55:07 crc kubenswrapper[4799]: I1129 04:55:07.900516 4799 scope.go:117] "RemoveContainer" containerID="99c5c08976d14d2fb8bfd60a883c0eaa07a2b4ec5ce6f9de4d8a5dd6e5c7212d" Nov 29 04:55:08 crc kubenswrapper[4799]: I1129 04:55:08.281749 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-sz8zj" Nov 29 04:55:08 crc kubenswrapper[4799]: I1129 04:55:08.443886 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e1b036a-a1f1-44ed-9823-b49199c5b4de-config-data\") pod \"3e1b036a-a1f1-44ed-9823-b49199c5b4de\" (UID: \"3e1b036a-a1f1-44ed-9823-b49199c5b4de\") " Nov 29 04:55:08 crc kubenswrapper[4799]: I1129 04:55:08.443974 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3e1b036a-a1f1-44ed-9823-b49199c5b4de-fernet-keys\") pod \"3e1b036a-a1f1-44ed-9823-b49199c5b4de\" (UID: \"3e1b036a-a1f1-44ed-9823-b49199c5b4de\") " Nov 29 04:55:08 crc kubenswrapper[4799]: I1129 04:55:08.444017 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kh29k\" (UniqueName: \"kubernetes.io/projected/3e1b036a-a1f1-44ed-9823-b49199c5b4de-kube-api-access-kh29k\") pod \"3e1b036a-a1f1-44ed-9823-b49199c5b4de\" (UID: \"3e1b036a-a1f1-44ed-9823-b49199c5b4de\") " Nov 29 04:55:08 crc kubenswrapper[4799]: I1129 04:55:08.444076 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3e1b036a-a1f1-44ed-9823-b49199c5b4de-credential-keys\") pod \"3e1b036a-a1f1-44ed-9823-b49199c5b4de\" (UID: \"3e1b036a-a1f1-44ed-9823-b49199c5b4de\") " Nov 29 04:55:08 crc kubenswrapper[4799]: I1129 04:55:08.444171 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e1b036a-a1f1-44ed-9823-b49199c5b4de-combined-ca-bundle\") pod \"3e1b036a-a1f1-44ed-9823-b49199c5b4de\" (UID: \"3e1b036a-a1f1-44ed-9823-b49199c5b4de\") " Nov 29 04:55:08 crc kubenswrapper[4799]: I1129 04:55:08.444248 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e1b036a-a1f1-44ed-9823-b49199c5b4de-scripts\") pod \"3e1b036a-a1f1-44ed-9823-b49199c5b4de\" (UID: \"3e1b036a-a1f1-44ed-9823-b49199c5b4de\") " Nov 29 04:55:08 crc kubenswrapper[4799]: I1129 04:55:08.452953 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e1b036a-a1f1-44ed-9823-b49199c5b4de-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "3e1b036a-a1f1-44ed-9823-b49199c5b4de" (UID: "3e1b036a-a1f1-44ed-9823-b49199c5b4de"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:55:08 crc kubenswrapper[4799]: I1129 04:55:08.454465 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e1b036a-a1f1-44ed-9823-b49199c5b4de-scripts" (OuterVolumeSpecName: "scripts") pod "3e1b036a-a1f1-44ed-9823-b49199c5b4de" (UID: "3e1b036a-a1f1-44ed-9823-b49199c5b4de"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:55:08 crc kubenswrapper[4799]: I1129 04:55:08.454757 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e1b036a-a1f1-44ed-9823-b49199c5b4de-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "3e1b036a-a1f1-44ed-9823-b49199c5b4de" (UID: "3e1b036a-a1f1-44ed-9823-b49199c5b4de"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:55:08 crc kubenswrapper[4799]: I1129 04:55:08.460672 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e1b036a-a1f1-44ed-9823-b49199c5b4de-kube-api-access-kh29k" (OuterVolumeSpecName: "kube-api-access-kh29k") pod "3e1b036a-a1f1-44ed-9823-b49199c5b4de" (UID: "3e1b036a-a1f1-44ed-9823-b49199c5b4de"). InnerVolumeSpecName "kube-api-access-kh29k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:55:08 crc kubenswrapper[4799]: I1129 04:55:08.474317 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e1b036a-a1f1-44ed-9823-b49199c5b4de-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3e1b036a-a1f1-44ed-9823-b49199c5b4de" (UID: "3e1b036a-a1f1-44ed-9823-b49199c5b4de"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:55:08 crc kubenswrapper[4799]: I1129 04:55:08.475024 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e1b036a-a1f1-44ed-9823-b49199c5b4de-config-data" (OuterVolumeSpecName: "config-data") pod "3e1b036a-a1f1-44ed-9823-b49199c5b4de" (UID: "3e1b036a-a1f1-44ed-9823-b49199c5b4de"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:55:08 crc kubenswrapper[4799]: I1129 04:55:08.546186 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e1b036a-a1f1-44ed-9823-b49199c5b4de-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:08 crc kubenswrapper[4799]: I1129 04:55:08.546609 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e1b036a-a1f1-44ed-9823-b49199c5b4de-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:08 crc kubenswrapper[4799]: I1129 04:55:08.546620 4799 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3e1b036a-a1f1-44ed-9823-b49199c5b4de-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:08 crc kubenswrapper[4799]: I1129 04:55:08.546635 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kh29k\" (UniqueName: \"kubernetes.io/projected/3e1b036a-a1f1-44ed-9823-b49199c5b4de-kube-api-access-kh29k\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:08 crc kubenswrapper[4799]: I1129 04:55:08.546650 4799 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3e1b036a-a1f1-44ed-9823-b49199c5b4de-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:08 crc kubenswrapper[4799]: I1129 04:55:08.546662 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e1b036a-a1f1-44ed-9823-b49199c5b4de-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:08 crc kubenswrapper[4799]: I1129 04:55:08.926179 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-sz8zj" Nov 29 04:55:08 crc kubenswrapper[4799]: I1129 04:55:08.926202 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-sz8zj" event={"ID":"3e1b036a-a1f1-44ed-9823-b49199c5b4de","Type":"ContainerDied","Data":"36fafcbe0ace23e99b5007d42af4eebf4548a5e8f190ea49d398d56790f2d4ba"} Nov 29 04:55:08 crc kubenswrapper[4799]: I1129 04:55:08.926242 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="36fafcbe0ace23e99b5007d42af4eebf4548a5e8f190ea49d398d56790f2d4ba" Nov 29 04:55:08 crc kubenswrapper[4799]: I1129 04:55:08.933439 4799 generic.go:334] "Generic (PLEG): container finished" podID="82f8e690-e10d-4852-bc45-beba0e5b666e" containerID="7d8eadeceea35e76b407cce3a32ef5e187792093f406f2a855c2100538b72f7b" exitCode=0 Nov 29 04:55:08 crc kubenswrapper[4799]: I1129 04:55:08.933499 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-dbv6d" event={"ID":"82f8e690-e10d-4852-bc45-beba0e5b666e","Type":"ContainerDied","Data":"7d8eadeceea35e76b407cce3a32ef5e187792093f406f2a855c2100538b72f7b"} Nov 29 04:55:09 crc kubenswrapper[4799]: I1129 04:55:09.396881 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-sz8zj"] Nov 29 04:55:09 crc kubenswrapper[4799]: I1129 04:55:09.402959 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-sz8zj"] Nov 29 04:55:09 crc kubenswrapper[4799]: I1129 04:55:09.487194 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-jtjzp"] Nov 29 04:55:09 crc kubenswrapper[4799]: E1129 04:55:09.488040 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e1b036a-a1f1-44ed-9823-b49199c5b4de" containerName="keystone-bootstrap" Nov 29 04:55:09 crc kubenswrapper[4799]: I1129 04:55:09.488187 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e1b036a-a1f1-44ed-9823-b49199c5b4de" containerName="keystone-bootstrap" Nov 29 04:55:09 crc kubenswrapper[4799]: E1129 04:55:09.488266 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64328f76-389c-43d8-9f70-e6888983268e" containerName="init" Nov 29 04:55:09 crc kubenswrapper[4799]: I1129 04:55:09.488328 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="64328f76-389c-43d8-9f70-e6888983268e" containerName="init" Nov 29 04:55:09 crc kubenswrapper[4799]: E1129 04:55:09.488395 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dabdbea-d929-4b73-85a4-b6a882967131" containerName="init" Nov 29 04:55:09 crc kubenswrapper[4799]: I1129 04:55:09.488486 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dabdbea-d929-4b73-85a4-b6a882967131" containerName="init" Nov 29 04:55:09 crc kubenswrapper[4799]: I1129 04:55:09.488803 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e1b036a-a1f1-44ed-9823-b49199c5b4de" containerName="keystone-bootstrap" Nov 29 04:55:09 crc kubenswrapper[4799]: I1129 04:55:09.488965 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="64328f76-389c-43d8-9f70-e6888983268e" containerName="init" Nov 29 04:55:09 crc kubenswrapper[4799]: I1129 04:55:09.489042 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="2dabdbea-d929-4b73-85a4-b6a882967131" containerName="init" Nov 29 04:55:09 crc kubenswrapper[4799]: I1129 04:55:09.489782 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jtjzp" Nov 29 04:55:09 crc kubenswrapper[4799]: I1129 04:55:09.492113 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-85vg6" Nov 29 04:55:09 crc kubenswrapper[4799]: I1129 04:55:09.493040 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 29 04:55:09 crc kubenswrapper[4799]: I1129 04:55:09.493267 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 29 04:55:09 crc kubenswrapper[4799]: I1129 04:55:09.493382 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 29 04:55:09 crc kubenswrapper[4799]: I1129 04:55:09.493594 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 29 04:55:09 crc kubenswrapper[4799]: I1129 04:55:09.497470 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-jtjzp"] Nov 29 04:55:09 crc kubenswrapper[4799]: I1129 04:55:09.681331 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96493a78-8bf7-49d4-85a0-748217f66262-scripts\") pod \"keystone-bootstrap-jtjzp\" (UID: \"96493a78-8bf7-49d4-85a0-748217f66262\") " pod="openstack/keystone-bootstrap-jtjzp" Nov 29 04:55:09 crc kubenswrapper[4799]: I1129 04:55:09.681935 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/96493a78-8bf7-49d4-85a0-748217f66262-fernet-keys\") pod \"keystone-bootstrap-jtjzp\" (UID: \"96493a78-8bf7-49d4-85a0-748217f66262\") " pod="openstack/keystone-bootstrap-jtjzp" Nov 29 04:55:09 crc kubenswrapper[4799]: I1129 04:55:09.682094 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cr2k8\" (UniqueName: \"kubernetes.io/projected/96493a78-8bf7-49d4-85a0-748217f66262-kube-api-access-cr2k8\") pod \"keystone-bootstrap-jtjzp\" (UID: \"96493a78-8bf7-49d4-85a0-748217f66262\") " pod="openstack/keystone-bootstrap-jtjzp" Nov 29 04:55:09 crc kubenswrapper[4799]: I1129 04:55:09.682140 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96493a78-8bf7-49d4-85a0-748217f66262-config-data\") pod \"keystone-bootstrap-jtjzp\" (UID: \"96493a78-8bf7-49d4-85a0-748217f66262\") " pod="openstack/keystone-bootstrap-jtjzp" Nov 29 04:55:09 crc kubenswrapper[4799]: I1129 04:55:09.682422 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/96493a78-8bf7-49d4-85a0-748217f66262-credential-keys\") pod \"keystone-bootstrap-jtjzp\" (UID: \"96493a78-8bf7-49d4-85a0-748217f66262\") " pod="openstack/keystone-bootstrap-jtjzp" Nov 29 04:55:09 crc kubenswrapper[4799]: I1129 04:55:09.682564 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96493a78-8bf7-49d4-85a0-748217f66262-combined-ca-bundle\") pod \"keystone-bootstrap-jtjzp\" (UID: \"96493a78-8bf7-49d4-85a0-748217f66262\") " pod="openstack/keystone-bootstrap-jtjzp" Nov 29 04:55:09 crc kubenswrapper[4799]: I1129 04:55:09.784212 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96493a78-8bf7-49d4-85a0-748217f66262-scripts\") pod \"keystone-bootstrap-jtjzp\" (UID: \"96493a78-8bf7-49d4-85a0-748217f66262\") " pod="openstack/keystone-bootstrap-jtjzp" Nov 29 04:55:09 crc kubenswrapper[4799]: I1129 04:55:09.784302 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/96493a78-8bf7-49d4-85a0-748217f66262-fernet-keys\") pod \"keystone-bootstrap-jtjzp\" (UID: \"96493a78-8bf7-49d4-85a0-748217f66262\") " pod="openstack/keystone-bootstrap-jtjzp" Nov 29 04:55:09 crc kubenswrapper[4799]: I1129 04:55:09.784360 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cr2k8\" (UniqueName: \"kubernetes.io/projected/96493a78-8bf7-49d4-85a0-748217f66262-kube-api-access-cr2k8\") pod \"keystone-bootstrap-jtjzp\" (UID: \"96493a78-8bf7-49d4-85a0-748217f66262\") " pod="openstack/keystone-bootstrap-jtjzp" Nov 29 04:55:09 crc kubenswrapper[4799]: I1129 04:55:09.784384 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96493a78-8bf7-49d4-85a0-748217f66262-config-data\") pod \"keystone-bootstrap-jtjzp\" (UID: \"96493a78-8bf7-49d4-85a0-748217f66262\") " pod="openstack/keystone-bootstrap-jtjzp" Nov 29 04:55:09 crc kubenswrapper[4799]: I1129 04:55:09.784467 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/96493a78-8bf7-49d4-85a0-748217f66262-credential-keys\") pod \"keystone-bootstrap-jtjzp\" (UID: \"96493a78-8bf7-49d4-85a0-748217f66262\") " pod="openstack/keystone-bootstrap-jtjzp" Nov 29 04:55:09 crc kubenswrapper[4799]: I1129 04:55:09.784692 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96493a78-8bf7-49d4-85a0-748217f66262-combined-ca-bundle\") pod \"keystone-bootstrap-jtjzp\" (UID: \"96493a78-8bf7-49d4-85a0-748217f66262\") " pod="openstack/keystone-bootstrap-jtjzp" Nov 29 04:55:09 crc kubenswrapper[4799]: I1129 04:55:09.792896 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/96493a78-8bf7-49d4-85a0-748217f66262-credential-keys\") pod \"keystone-bootstrap-jtjzp\" (UID: \"96493a78-8bf7-49d4-85a0-748217f66262\") " pod="openstack/keystone-bootstrap-jtjzp" Nov 29 04:55:09 crc kubenswrapper[4799]: I1129 04:55:09.793543 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/96493a78-8bf7-49d4-85a0-748217f66262-fernet-keys\") pod \"keystone-bootstrap-jtjzp\" (UID: \"96493a78-8bf7-49d4-85a0-748217f66262\") " pod="openstack/keystone-bootstrap-jtjzp" Nov 29 04:55:09 crc kubenswrapper[4799]: I1129 04:55:09.799034 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96493a78-8bf7-49d4-85a0-748217f66262-config-data\") pod \"keystone-bootstrap-jtjzp\" (UID: \"96493a78-8bf7-49d4-85a0-748217f66262\") " pod="openstack/keystone-bootstrap-jtjzp" Nov 29 04:55:09 crc kubenswrapper[4799]: I1129 04:55:09.799719 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96493a78-8bf7-49d4-85a0-748217f66262-scripts\") pod \"keystone-bootstrap-jtjzp\" (UID: \"96493a78-8bf7-49d4-85a0-748217f66262\") " pod="openstack/keystone-bootstrap-jtjzp" Nov 29 04:55:09 crc kubenswrapper[4799]: I1129 04:55:09.805629 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96493a78-8bf7-49d4-85a0-748217f66262-combined-ca-bundle\") pod \"keystone-bootstrap-jtjzp\" (UID: \"96493a78-8bf7-49d4-85a0-748217f66262\") " pod="openstack/keystone-bootstrap-jtjzp" Nov 29 04:55:09 crc kubenswrapper[4799]: I1129 04:55:09.806852 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cr2k8\" (UniqueName: \"kubernetes.io/projected/96493a78-8bf7-49d4-85a0-748217f66262-kube-api-access-cr2k8\") pod \"keystone-bootstrap-jtjzp\" (UID: \"96493a78-8bf7-49d4-85a0-748217f66262\") " pod="openstack/keystone-bootstrap-jtjzp" Nov 29 04:55:09 crc kubenswrapper[4799]: I1129 04:55:09.813888 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jtjzp" Nov 29 04:55:10 crc kubenswrapper[4799]: I1129 04:55:10.673915 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e1b036a-a1f1-44ed-9823-b49199c5b4de" path="/var/lib/kubelet/pods/3e1b036a-a1f1-44ed-9823-b49199c5b4de/volumes" Nov 29 04:55:15 crc kubenswrapper[4799]: I1129 04:55:15.985998 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-86db49b7ff-dbv6d" podUID="82f8e690-e10d-4852-bc45-beba0e5b666e" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.109:5353: i/o timeout" Nov 29 04:55:20 crc kubenswrapper[4799]: I1129 04:55:20.986899 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-86db49b7ff-dbv6d" podUID="82f8e690-e10d-4852-bc45-beba0e5b666e" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.109:5353: i/o timeout" Nov 29 04:55:25 crc kubenswrapper[4799]: I1129 04:55:25.987867 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-86db49b7ff-dbv6d" podUID="82f8e690-e10d-4852-bc45-beba0e5b666e" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.109:5353: i/o timeout" Nov 29 04:55:25 crc kubenswrapper[4799]: I1129 04:55:25.988650 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-dbv6d" Nov 29 04:55:27 crc kubenswrapper[4799]: I1129 04:55:27.293584 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-dbv6d" Nov 29 04:55:27 crc kubenswrapper[4799]: I1129 04:55:27.414498 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/82f8e690-e10d-4852-bc45-beba0e5b666e-ovsdbserver-nb\") pod \"82f8e690-e10d-4852-bc45-beba0e5b666e\" (UID: \"82f8e690-e10d-4852-bc45-beba0e5b666e\") " Nov 29 04:55:27 crc kubenswrapper[4799]: I1129 04:55:27.414598 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/82f8e690-e10d-4852-bc45-beba0e5b666e-dns-svc\") pod \"82f8e690-e10d-4852-bc45-beba0e5b666e\" (UID: \"82f8e690-e10d-4852-bc45-beba0e5b666e\") " Nov 29 04:55:27 crc kubenswrapper[4799]: I1129 04:55:27.414718 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82f8e690-e10d-4852-bc45-beba0e5b666e-config\") pod \"82f8e690-e10d-4852-bc45-beba0e5b666e\" (UID: \"82f8e690-e10d-4852-bc45-beba0e5b666e\") " Nov 29 04:55:27 crc kubenswrapper[4799]: I1129 04:55:27.414880 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-njqgn\" (UniqueName: \"kubernetes.io/projected/82f8e690-e10d-4852-bc45-beba0e5b666e-kube-api-access-njqgn\") pod \"82f8e690-e10d-4852-bc45-beba0e5b666e\" (UID: \"82f8e690-e10d-4852-bc45-beba0e5b666e\") " Nov 29 04:55:27 crc kubenswrapper[4799]: I1129 04:55:27.415547 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/82f8e690-e10d-4852-bc45-beba0e5b666e-ovsdbserver-sb\") pod \"82f8e690-e10d-4852-bc45-beba0e5b666e\" (UID: \"82f8e690-e10d-4852-bc45-beba0e5b666e\") " Nov 29 04:55:27 crc kubenswrapper[4799]: I1129 04:55:27.422799 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82f8e690-e10d-4852-bc45-beba0e5b666e-kube-api-access-njqgn" (OuterVolumeSpecName: "kube-api-access-njqgn") pod "82f8e690-e10d-4852-bc45-beba0e5b666e" (UID: "82f8e690-e10d-4852-bc45-beba0e5b666e"). InnerVolumeSpecName "kube-api-access-njqgn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:55:27 crc kubenswrapper[4799]: E1129 04:55:27.460857 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/82f8e690-e10d-4852-bc45-beba0e5b666e-ovsdbserver-nb podName:82f8e690-e10d-4852-bc45-beba0e5b666e nodeName:}" failed. No retries permitted until 2025-11-29 04:55:27.960730808 +0000 UTC m=+1003.603661208 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "ovsdbserver-nb" (UniqueName: "kubernetes.io/configmap/82f8e690-e10d-4852-bc45-beba0e5b666e-ovsdbserver-nb") pod "82f8e690-e10d-4852-bc45-beba0e5b666e" (UID: "82f8e690-e10d-4852-bc45-beba0e5b666e") : error deleting /var/lib/kubelet/pods/82f8e690-e10d-4852-bc45-beba0e5b666e/volume-subpaths: remove /var/lib/kubelet/pods/82f8e690-e10d-4852-bc45-beba0e5b666e/volume-subpaths: no such file or directory Nov 29 04:55:27 crc kubenswrapper[4799]: I1129 04:55:27.461201 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82f8e690-e10d-4852-bc45-beba0e5b666e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "82f8e690-e10d-4852-bc45-beba0e5b666e" (UID: "82f8e690-e10d-4852-bc45-beba0e5b666e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:55:27 crc kubenswrapper[4799]: I1129 04:55:27.461857 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82f8e690-e10d-4852-bc45-beba0e5b666e-config" (OuterVolumeSpecName: "config") pod "82f8e690-e10d-4852-bc45-beba0e5b666e" (UID: "82f8e690-e10d-4852-bc45-beba0e5b666e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:55:27 crc kubenswrapper[4799]: I1129 04:55:27.517863 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-njqgn\" (UniqueName: \"kubernetes.io/projected/82f8e690-e10d-4852-bc45-beba0e5b666e-kube-api-access-njqgn\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:27 crc kubenswrapper[4799]: I1129 04:55:27.517905 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/82f8e690-e10d-4852-bc45-beba0e5b666e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:27 crc kubenswrapper[4799]: I1129 04:55:27.517917 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82f8e690-e10d-4852-bc45-beba0e5b666e-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:27 crc kubenswrapper[4799]: I1129 04:55:27.554979 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82f8e690-e10d-4852-bc45-beba0e5b666e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "82f8e690-e10d-4852-bc45-beba0e5b666e" (UID: "82f8e690-e10d-4852-bc45-beba0e5b666e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:55:27 crc kubenswrapper[4799]: E1129 04:55:27.591106 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Nov 29 04:55:27 crc kubenswrapper[4799]: E1129 04:55:27.591326 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mgq57,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-gcfzn_openstack(8370f287-fa77-4ef5-b929-01a4cf8c598c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 29 04:55:27 crc kubenswrapper[4799]: E1129 04:55:27.593251 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-gcfzn" podUID="8370f287-fa77-4ef5-b929-01a4cf8c598c" Nov 29 04:55:27 crc kubenswrapper[4799]: I1129 04:55:27.620366 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/82f8e690-e10d-4852-bc45-beba0e5b666e-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:28 crc kubenswrapper[4799]: I1129 04:55:28.029528 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/82f8e690-e10d-4852-bc45-beba0e5b666e-ovsdbserver-nb\") pod \"82f8e690-e10d-4852-bc45-beba0e5b666e\" (UID: \"82f8e690-e10d-4852-bc45-beba0e5b666e\") " Nov 29 04:55:28 crc kubenswrapper[4799]: I1129 04:55:28.030353 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82f8e690-e10d-4852-bc45-beba0e5b666e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "82f8e690-e10d-4852-bc45-beba0e5b666e" (UID: "82f8e690-e10d-4852-bc45-beba0e5b666e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:55:28 crc kubenswrapper[4799]: I1129 04:55:28.131690 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/82f8e690-e10d-4852-bc45-beba0e5b666e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:28 crc kubenswrapper[4799]: I1129 04:55:28.138568 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-dbv6d" Nov 29 04:55:28 crc kubenswrapper[4799]: I1129 04:55:28.138549 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-dbv6d" event={"ID":"82f8e690-e10d-4852-bc45-beba0e5b666e","Type":"ContainerDied","Data":"b8fa8d7f083f296493e244bc00d86d87ebbd00fe65007fdd24d4a61c202b0354"} Nov 29 04:55:28 crc kubenswrapper[4799]: E1129 04:55:28.140212 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-gcfzn" podUID="8370f287-fa77-4ef5-b929-01a4cf8c598c" Nov 29 04:55:28 crc kubenswrapper[4799]: I1129 04:55:28.186646 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-dbv6d"] Nov 29 04:55:28 crc kubenswrapper[4799]: I1129 04:55:28.196414 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-dbv6d"] Nov 29 04:55:28 crc kubenswrapper[4799]: I1129 04:55:28.565588 4799 scope.go:117] "RemoveContainer" containerID="7d8eadeceea35e76b407cce3a32ef5e187792093f406f2a855c2100538b72f7b" Nov 29 04:55:28 crc kubenswrapper[4799]: I1129 04:55:28.592693 4799 scope.go:117] "RemoveContainer" containerID="e0521ca520400ce0ca8b2930b538fc5edee5a488dc7c80d203302043be8fb969" Nov 29 04:55:28 crc kubenswrapper[4799]: E1129 04:55:28.617419 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Nov 29 04:55:28 crc kubenswrapper[4799]: E1129 04:55:28.617597 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-shprr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-btd8j_openstack(515c81c7-10e3-4724-aafd-42431946c2c2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 29 04:55:28 crc kubenswrapper[4799]: E1129 04:55:28.619982 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-btd8j" podUID="515c81c7-10e3-4724-aafd-42431946c2c2" Nov 29 04:55:28 crc kubenswrapper[4799]: I1129 04:55:28.692249 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82f8e690-e10d-4852-bc45-beba0e5b666e" path="/var/lib/kubelet/pods/82f8e690-e10d-4852-bc45-beba0e5b666e/volumes" Nov 29 04:55:29 crc kubenswrapper[4799]: I1129 04:55:29.019894 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-jtjzp"] Nov 29 04:55:29 crc kubenswrapper[4799]: I1129 04:55:29.151264 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" event={"ID":"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8","Type":"ContainerStarted","Data":"0bb5e3708c6fb82d15eb8dc7170b3147f6cdc27661c141f7e521f399422452a4"} Nov 29 04:55:29 crc kubenswrapper[4799]: I1129 04:55:29.159455 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-6cz7h" event={"ID":"57a0c30f-28ae-466a-bc63-59ad7b6afd51","Type":"ContainerStarted","Data":"a1c1ce346c940a7f505e01f058700d6541525b7df547eeb2db3011375df3404d"} Nov 29 04:55:29 crc kubenswrapper[4799]: I1129 04:55:29.161981 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jtjzp" event={"ID":"96493a78-8bf7-49d4-85a0-748217f66262","Type":"ContainerStarted","Data":"69264827b6759fc61402a802e0708fffc7501874a55274da76e1237d6b87164c"} Nov 29 04:55:29 crc kubenswrapper[4799]: I1129 04:55:29.172720 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"82c42edf-7b25-411f-b00f-95ed3e74bede","Type":"ContainerStarted","Data":"19f81f19bd1e89c5024a9fd7c16e1c64a6b91117546b1cfd8f819f63de7fdfb1"} Nov 29 04:55:29 crc kubenswrapper[4799]: E1129 04:55:29.176535 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-btd8j" podUID="515c81c7-10e3-4724-aafd-42431946c2c2" Nov 29 04:55:29 crc kubenswrapper[4799]: I1129 04:55:29.216691 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-6cz7h" podStartSLOduration=2.037973474 podStartE2EDuration="32.216668995s" podCreationTimestamp="2025-11-29 04:54:57 +0000 UTC" firstStartedPulling="2025-11-29 04:54:58.399885473 +0000 UTC m=+974.042815873" lastFinishedPulling="2025-11-29 04:55:28.578580994 +0000 UTC m=+1004.221511394" observedRunningTime="2025-11-29 04:55:29.213430395 +0000 UTC m=+1004.856360795" watchObservedRunningTime="2025-11-29 04:55:29.216668995 +0000 UTC m=+1004.859599395" Nov 29 04:55:30 crc kubenswrapper[4799]: I1129 04:55:30.192126 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jtjzp" event={"ID":"96493a78-8bf7-49d4-85a0-748217f66262","Type":"ContainerStarted","Data":"0786a8d8316759000427db9e93b6c94a09927610f0b98a13528749ed4a57ae4c"} Nov 29 04:55:30 crc kubenswrapper[4799]: I1129 04:55:30.197318 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"82c42edf-7b25-411f-b00f-95ed3e74bede","Type":"ContainerStarted","Data":"029c5b8b863f82633a82b4acc01bab9cc121c4cc9a483238bf2cf3a6bd02cb66"} Nov 29 04:55:30 crc kubenswrapper[4799]: I1129 04:55:30.218779 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-jtjzp" podStartSLOduration=21.218755088 podStartE2EDuration="21.218755088s" podCreationTimestamp="2025-11-29 04:55:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:55:30.21314141 +0000 UTC m=+1005.856071810" watchObservedRunningTime="2025-11-29 04:55:30.218755088 +0000 UTC m=+1005.861685488" Nov 29 04:55:30 crc kubenswrapper[4799]: I1129 04:55:30.989448 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-86db49b7ff-dbv6d" podUID="82f8e690-e10d-4852-bc45-beba0e5b666e" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.109:5353: i/o timeout" Nov 29 04:55:32 crc kubenswrapper[4799]: I1129 04:55:32.219077 4799 generic.go:334] "Generic (PLEG): container finished" podID="57a0c30f-28ae-466a-bc63-59ad7b6afd51" containerID="a1c1ce346c940a7f505e01f058700d6541525b7df547eeb2db3011375df3404d" exitCode=0 Nov 29 04:55:32 crc kubenswrapper[4799]: I1129 04:55:32.219125 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-6cz7h" event={"ID":"57a0c30f-28ae-466a-bc63-59ad7b6afd51","Type":"ContainerDied","Data":"a1c1ce346c940a7f505e01f058700d6541525b7df547eeb2db3011375df3404d"} Nov 29 04:55:33 crc kubenswrapper[4799]: I1129 04:55:33.560225 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-6cz7h" Nov 29 04:55:33 crc kubenswrapper[4799]: I1129 04:55:33.743921 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/57a0c30f-28ae-466a-bc63-59ad7b6afd51-logs\") pod \"57a0c30f-28ae-466a-bc63-59ad7b6afd51\" (UID: \"57a0c30f-28ae-466a-bc63-59ad7b6afd51\") " Nov 29 04:55:33 crc kubenswrapper[4799]: I1129 04:55:33.743989 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zjqqw\" (UniqueName: \"kubernetes.io/projected/57a0c30f-28ae-466a-bc63-59ad7b6afd51-kube-api-access-zjqqw\") pod \"57a0c30f-28ae-466a-bc63-59ad7b6afd51\" (UID: \"57a0c30f-28ae-466a-bc63-59ad7b6afd51\") " Nov 29 04:55:33 crc kubenswrapper[4799]: I1129 04:55:33.744018 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57a0c30f-28ae-466a-bc63-59ad7b6afd51-combined-ca-bundle\") pod \"57a0c30f-28ae-466a-bc63-59ad7b6afd51\" (UID: \"57a0c30f-28ae-466a-bc63-59ad7b6afd51\") " Nov 29 04:55:33 crc kubenswrapper[4799]: I1129 04:55:33.744141 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57a0c30f-28ae-466a-bc63-59ad7b6afd51-scripts\") pod \"57a0c30f-28ae-466a-bc63-59ad7b6afd51\" (UID: \"57a0c30f-28ae-466a-bc63-59ad7b6afd51\") " Nov 29 04:55:33 crc kubenswrapper[4799]: I1129 04:55:33.744192 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57a0c30f-28ae-466a-bc63-59ad7b6afd51-config-data\") pod \"57a0c30f-28ae-466a-bc63-59ad7b6afd51\" (UID: \"57a0c30f-28ae-466a-bc63-59ad7b6afd51\") " Nov 29 04:55:33 crc kubenswrapper[4799]: I1129 04:55:33.744756 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a0c30f-28ae-466a-bc63-59ad7b6afd51-logs" (OuterVolumeSpecName: "logs") pod "57a0c30f-28ae-466a-bc63-59ad7b6afd51" (UID: "57a0c30f-28ae-466a-bc63-59ad7b6afd51"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:55:33 crc kubenswrapper[4799]: I1129 04:55:33.760806 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a0c30f-28ae-466a-bc63-59ad7b6afd51-kube-api-access-zjqqw" (OuterVolumeSpecName: "kube-api-access-zjqqw") pod "57a0c30f-28ae-466a-bc63-59ad7b6afd51" (UID: "57a0c30f-28ae-466a-bc63-59ad7b6afd51"). InnerVolumeSpecName "kube-api-access-zjqqw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:55:33 crc kubenswrapper[4799]: I1129 04:55:33.761946 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57a0c30f-28ae-466a-bc63-59ad7b6afd51-scripts" (OuterVolumeSpecName: "scripts") pod "57a0c30f-28ae-466a-bc63-59ad7b6afd51" (UID: "57a0c30f-28ae-466a-bc63-59ad7b6afd51"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:55:33 crc kubenswrapper[4799]: I1129 04:55:33.772485 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57a0c30f-28ae-466a-bc63-59ad7b6afd51-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "57a0c30f-28ae-466a-bc63-59ad7b6afd51" (UID: "57a0c30f-28ae-466a-bc63-59ad7b6afd51"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:55:33 crc kubenswrapper[4799]: I1129 04:55:33.772926 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57a0c30f-28ae-466a-bc63-59ad7b6afd51-config-data" (OuterVolumeSpecName: "config-data") pod "57a0c30f-28ae-466a-bc63-59ad7b6afd51" (UID: "57a0c30f-28ae-466a-bc63-59ad7b6afd51"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:55:33 crc kubenswrapper[4799]: I1129 04:55:33.847277 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57a0c30f-28ae-466a-bc63-59ad7b6afd51-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:33 crc kubenswrapper[4799]: I1129 04:55:33.847308 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57a0c30f-28ae-466a-bc63-59ad7b6afd51-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:33 crc kubenswrapper[4799]: I1129 04:55:33.847318 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/57a0c30f-28ae-466a-bc63-59ad7b6afd51-logs\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:33 crc kubenswrapper[4799]: I1129 04:55:33.847326 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zjqqw\" (UniqueName: \"kubernetes.io/projected/57a0c30f-28ae-466a-bc63-59ad7b6afd51-kube-api-access-zjqqw\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:33 crc kubenswrapper[4799]: I1129 04:55:33.847337 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57a0c30f-28ae-466a-bc63-59ad7b6afd51-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:34 crc kubenswrapper[4799]: I1129 04:55:34.244708 4799 generic.go:334] "Generic (PLEG): container finished" podID="96493a78-8bf7-49d4-85a0-748217f66262" containerID="0786a8d8316759000427db9e93b6c94a09927610f0b98a13528749ed4a57ae4c" exitCode=0 Nov 29 04:55:34 crc kubenswrapper[4799]: I1129 04:55:34.244829 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jtjzp" event={"ID":"96493a78-8bf7-49d4-85a0-748217f66262","Type":"ContainerDied","Data":"0786a8d8316759000427db9e93b6c94a09927610f0b98a13528749ed4a57ae4c"} Nov 29 04:55:34 crc kubenswrapper[4799]: I1129 04:55:34.246173 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-6cz7h" event={"ID":"57a0c30f-28ae-466a-bc63-59ad7b6afd51","Type":"ContainerDied","Data":"1474247c5094f84feb0e2529060061812c7fe23b552ef19e6ff9efa77463600f"} Nov 29 04:55:34 crc kubenswrapper[4799]: I1129 04:55:34.246222 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1474247c5094f84feb0e2529060061812c7fe23b552ef19e6ff9efa77463600f" Nov 29 04:55:34 crc kubenswrapper[4799]: I1129 04:55:34.246275 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-6cz7h" Nov 29 04:55:34 crc kubenswrapper[4799]: I1129 04:55:34.353705 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-dc694ffdd-sm2s5"] Nov 29 04:55:34 crc kubenswrapper[4799]: E1129 04:55:34.354271 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82f8e690-e10d-4852-bc45-beba0e5b666e" containerName="init" Nov 29 04:55:34 crc kubenswrapper[4799]: I1129 04:55:34.354292 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="82f8e690-e10d-4852-bc45-beba0e5b666e" containerName="init" Nov 29 04:55:34 crc kubenswrapper[4799]: E1129 04:55:34.354328 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57a0c30f-28ae-466a-bc63-59ad7b6afd51" containerName="placement-db-sync" Nov 29 04:55:34 crc kubenswrapper[4799]: I1129 04:55:34.354335 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="57a0c30f-28ae-466a-bc63-59ad7b6afd51" containerName="placement-db-sync" Nov 29 04:55:34 crc kubenswrapper[4799]: E1129 04:55:34.354359 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82f8e690-e10d-4852-bc45-beba0e5b666e" containerName="dnsmasq-dns" Nov 29 04:55:34 crc kubenswrapper[4799]: I1129 04:55:34.354364 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="82f8e690-e10d-4852-bc45-beba0e5b666e" containerName="dnsmasq-dns" Nov 29 04:55:34 crc kubenswrapper[4799]: I1129 04:55:34.354555 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="82f8e690-e10d-4852-bc45-beba0e5b666e" containerName="dnsmasq-dns" Nov 29 04:55:34 crc kubenswrapper[4799]: I1129 04:55:34.354576 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="57a0c30f-28ae-466a-bc63-59ad7b6afd51" containerName="placement-db-sync" Nov 29 04:55:34 crc kubenswrapper[4799]: I1129 04:55:34.355666 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-dc694ffdd-sm2s5" Nov 29 04:55:34 crc kubenswrapper[4799]: I1129 04:55:34.358215 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-mkdff" Nov 29 04:55:34 crc kubenswrapper[4799]: I1129 04:55:34.358435 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 29 04:55:34 crc kubenswrapper[4799]: I1129 04:55:34.360566 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Nov 29 04:55:34 crc kubenswrapper[4799]: I1129 04:55:34.360953 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Nov 29 04:55:34 crc kubenswrapper[4799]: I1129 04:55:34.370926 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-dc694ffdd-sm2s5"] Nov 29 04:55:34 crc kubenswrapper[4799]: I1129 04:55:34.379419 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 29 04:55:34 crc kubenswrapper[4799]: I1129 04:55:34.459803 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8af48d85-03bf-4d5f-a677-49bc96f0dfee-public-tls-certs\") pod \"placement-dc694ffdd-sm2s5\" (UID: \"8af48d85-03bf-4d5f-a677-49bc96f0dfee\") " pod="openstack/placement-dc694ffdd-sm2s5" Nov 29 04:55:34 crc kubenswrapper[4799]: I1129 04:55:34.459878 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8af48d85-03bf-4d5f-a677-49bc96f0dfee-config-data\") pod \"placement-dc694ffdd-sm2s5\" (UID: \"8af48d85-03bf-4d5f-a677-49bc96f0dfee\") " pod="openstack/placement-dc694ffdd-sm2s5" Nov 29 04:55:34 crc kubenswrapper[4799]: I1129 04:55:34.459924 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvtpk\" (UniqueName: \"kubernetes.io/projected/8af48d85-03bf-4d5f-a677-49bc96f0dfee-kube-api-access-pvtpk\") pod \"placement-dc694ffdd-sm2s5\" (UID: \"8af48d85-03bf-4d5f-a677-49bc96f0dfee\") " pod="openstack/placement-dc694ffdd-sm2s5" Nov 29 04:55:34 crc kubenswrapper[4799]: I1129 04:55:34.460130 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8af48d85-03bf-4d5f-a677-49bc96f0dfee-combined-ca-bundle\") pod \"placement-dc694ffdd-sm2s5\" (UID: \"8af48d85-03bf-4d5f-a677-49bc96f0dfee\") " pod="openstack/placement-dc694ffdd-sm2s5" Nov 29 04:55:34 crc kubenswrapper[4799]: I1129 04:55:34.460266 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8af48d85-03bf-4d5f-a677-49bc96f0dfee-logs\") pod \"placement-dc694ffdd-sm2s5\" (UID: \"8af48d85-03bf-4d5f-a677-49bc96f0dfee\") " pod="openstack/placement-dc694ffdd-sm2s5" Nov 29 04:55:34 crc kubenswrapper[4799]: I1129 04:55:34.460444 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8af48d85-03bf-4d5f-a677-49bc96f0dfee-internal-tls-certs\") pod \"placement-dc694ffdd-sm2s5\" (UID: \"8af48d85-03bf-4d5f-a677-49bc96f0dfee\") " pod="openstack/placement-dc694ffdd-sm2s5" Nov 29 04:55:34 crc kubenswrapper[4799]: I1129 04:55:34.460490 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8af48d85-03bf-4d5f-a677-49bc96f0dfee-scripts\") pod \"placement-dc694ffdd-sm2s5\" (UID: \"8af48d85-03bf-4d5f-a677-49bc96f0dfee\") " pod="openstack/placement-dc694ffdd-sm2s5" Nov 29 04:55:34 crc kubenswrapper[4799]: I1129 04:55:34.562813 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvtpk\" (UniqueName: \"kubernetes.io/projected/8af48d85-03bf-4d5f-a677-49bc96f0dfee-kube-api-access-pvtpk\") pod \"placement-dc694ffdd-sm2s5\" (UID: \"8af48d85-03bf-4d5f-a677-49bc96f0dfee\") " pod="openstack/placement-dc694ffdd-sm2s5" Nov 29 04:55:34 crc kubenswrapper[4799]: I1129 04:55:34.562903 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8af48d85-03bf-4d5f-a677-49bc96f0dfee-combined-ca-bundle\") pod \"placement-dc694ffdd-sm2s5\" (UID: \"8af48d85-03bf-4d5f-a677-49bc96f0dfee\") " pod="openstack/placement-dc694ffdd-sm2s5" Nov 29 04:55:34 crc kubenswrapper[4799]: I1129 04:55:34.562970 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8af48d85-03bf-4d5f-a677-49bc96f0dfee-logs\") pod \"placement-dc694ffdd-sm2s5\" (UID: \"8af48d85-03bf-4d5f-a677-49bc96f0dfee\") " pod="openstack/placement-dc694ffdd-sm2s5" Nov 29 04:55:34 crc kubenswrapper[4799]: I1129 04:55:34.563016 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8af48d85-03bf-4d5f-a677-49bc96f0dfee-internal-tls-certs\") pod \"placement-dc694ffdd-sm2s5\" (UID: \"8af48d85-03bf-4d5f-a677-49bc96f0dfee\") " pod="openstack/placement-dc694ffdd-sm2s5" Nov 29 04:55:34 crc kubenswrapper[4799]: I1129 04:55:34.563043 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8af48d85-03bf-4d5f-a677-49bc96f0dfee-scripts\") pod \"placement-dc694ffdd-sm2s5\" (UID: \"8af48d85-03bf-4d5f-a677-49bc96f0dfee\") " pod="openstack/placement-dc694ffdd-sm2s5" Nov 29 04:55:34 crc kubenswrapper[4799]: I1129 04:55:34.563116 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8af48d85-03bf-4d5f-a677-49bc96f0dfee-public-tls-certs\") pod \"placement-dc694ffdd-sm2s5\" (UID: \"8af48d85-03bf-4d5f-a677-49bc96f0dfee\") " pod="openstack/placement-dc694ffdd-sm2s5" Nov 29 04:55:34 crc kubenswrapper[4799]: I1129 04:55:34.563154 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8af48d85-03bf-4d5f-a677-49bc96f0dfee-config-data\") pod \"placement-dc694ffdd-sm2s5\" (UID: \"8af48d85-03bf-4d5f-a677-49bc96f0dfee\") " pod="openstack/placement-dc694ffdd-sm2s5" Nov 29 04:55:34 crc kubenswrapper[4799]: I1129 04:55:34.569071 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8af48d85-03bf-4d5f-a677-49bc96f0dfee-logs\") pod \"placement-dc694ffdd-sm2s5\" (UID: \"8af48d85-03bf-4d5f-a677-49bc96f0dfee\") " pod="openstack/placement-dc694ffdd-sm2s5" Nov 29 04:55:34 crc kubenswrapper[4799]: I1129 04:55:34.575773 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8af48d85-03bf-4d5f-a677-49bc96f0dfee-public-tls-certs\") pod \"placement-dc694ffdd-sm2s5\" (UID: \"8af48d85-03bf-4d5f-a677-49bc96f0dfee\") " pod="openstack/placement-dc694ffdd-sm2s5" Nov 29 04:55:34 crc kubenswrapper[4799]: I1129 04:55:34.580054 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8af48d85-03bf-4d5f-a677-49bc96f0dfee-combined-ca-bundle\") pod \"placement-dc694ffdd-sm2s5\" (UID: \"8af48d85-03bf-4d5f-a677-49bc96f0dfee\") " pod="openstack/placement-dc694ffdd-sm2s5" Nov 29 04:55:34 crc kubenswrapper[4799]: I1129 04:55:34.593437 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8af48d85-03bf-4d5f-a677-49bc96f0dfee-internal-tls-certs\") pod \"placement-dc694ffdd-sm2s5\" (UID: \"8af48d85-03bf-4d5f-a677-49bc96f0dfee\") " pod="openstack/placement-dc694ffdd-sm2s5" Nov 29 04:55:34 crc kubenswrapper[4799]: I1129 04:55:34.594142 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8af48d85-03bf-4d5f-a677-49bc96f0dfee-config-data\") pod \"placement-dc694ffdd-sm2s5\" (UID: \"8af48d85-03bf-4d5f-a677-49bc96f0dfee\") " pod="openstack/placement-dc694ffdd-sm2s5" Nov 29 04:55:34 crc kubenswrapper[4799]: I1129 04:55:34.608529 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8af48d85-03bf-4d5f-a677-49bc96f0dfee-scripts\") pod \"placement-dc694ffdd-sm2s5\" (UID: \"8af48d85-03bf-4d5f-a677-49bc96f0dfee\") " pod="openstack/placement-dc694ffdd-sm2s5" Nov 29 04:55:34 crc kubenswrapper[4799]: I1129 04:55:34.612172 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvtpk\" (UniqueName: \"kubernetes.io/projected/8af48d85-03bf-4d5f-a677-49bc96f0dfee-kube-api-access-pvtpk\") pod \"placement-dc694ffdd-sm2s5\" (UID: \"8af48d85-03bf-4d5f-a677-49bc96f0dfee\") " pod="openstack/placement-dc694ffdd-sm2s5" Nov 29 04:55:34 crc kubenswrapper[4799]: I1129 04:55:34.696730 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-dc694ffdd-sm2s5" Nov 29 04:55:35 crc kubenswrapper[4799]: I1129 04:55:35.632474 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jtjzp" Nov 29 04:55:35 crc kubenswrapper[4799]: I1129 04:55:35.790726 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cr2k8\" (UniqueName: \"kubernetes.io/projected/96493a78-8bf7-49d4-85a0-748217f66262-kube-api-access-cr2k8\") pod \"96493a78-8bf7-49d4-85a0-748217f66262\" (UID: \"96493a78-8bf7-49d4-85a0-748217f66262\") " Nov 29 04:55:35 crc kubenswrapper[4799]: I1129 04:55:35.790902 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96493a78-8bf7-49d4-85a0-748217f66262-config-data\") pod \"96493a78-8bf7-49d4-85a0-748217f66262\" (UID: \"96493a78-8bf7-49d4-85a0-748217f66262\") " Nov 29 04:55:35 crc kubenswrapper[4799]: I1129 04:55:35.790976 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/96493a78-8bf7-49d4-85a0-748217f66262-fernet-keys\") pod \"96493a78-8bf7-49d4-85a0-748217f66262\" (UID: \"96493a78-8bf7-49d4-85a0-748217f66262\") " Nov 29 04:55:35 crc kubenswrapper[4799]: I1129 04:55:35.791036 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96493a78-8bf7-49d4-85a0-748217f66262-combined-ca-bundle\") pod \"96493a78-8bf7-49d4-85a0-748217f66262\" (UID: \"96493a78-8bf7-49d4-85a0-748217f66262\") " Nov 29 04:55:35 crc kubenswrapper[4799]: I1129 04:55:35.791089 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/96493a78-8bf7-49d4-85a0-748217f66262-credential-keys\") pod \"96493a78-8bf7-49d4-85a0-748217f66262\" (UID: \"96493a78-8bf7-49d4-85a0-748217f66262\") " Nov 29 04:55:35 crc kubenswrapper[4799]: I1129 04:55:35.791152 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96493a78-8bf7-49d4-85a0-748217f66262-scripts\") pod \"96493a78-8bf7-49d4-85a0-748217f66262\" (UID: \"96493a78-8bf7-49d4-85a0-748217f66262\") " Nov 29 04:55:35 crc kubenswrapper[4799]: I1129 04:55:35.803342 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96493a78-8bf7-49d4-85a0-748217f66262-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "96493a78-8bf7-49d4-85a0-748217f66262" (UID: "96493a78-8bf7-49d4-85a0-748217f66262"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:55:35 crc kubenswrapper[4799]: I1129 04:55:35.803368 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96493a78-8bf7-49d4-85a0-748217f66262-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "96493a78-8bf7-49d4-85a0-748217f66262" (UID: "96493a78-8bf7-49d4-85a0-748217f66262"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:55:35 crc kubenswrapper[4799]: I1129 04:55:35.803503 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96493a78-8bf7-49d4-85a0-748217f66262-scripts" (OuterVolumeSpecName: "scripts") pod "96493a78-8bf7-49d4-85a0-748217f66262" (UID: "96493a78-8bf7-49d4-85a0-748217f66262"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:55:35 crc kubenswrapper[4799]: I1129 04:55:35.806017 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96493a78-8bf7-49d4-85a0-748217f66262-kube-api-access-cr2k8" (OuterVolumeSpecName: "kube-api-access-cr2k8") pod "96493a78-8bf7-49d4-85a0-748217f66262" (UID: "96493a78-8bf7-49d4-85a0-748217f66262"). InnerVolumeSpecName "kube-api-access-cr2k8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:55:35 crc kubenswrapper[4799]: I1129 04:55:35.829017 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96493a78-8bf7-49d4-85a0-748217f66262-config-data" (OuterVolumeSpecName: "config-data") pod "96493a78-8bf7-49d4-85a0-748217f66262" (UID: "96493a78-8bf7-49d4-85a0-748217f66262"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:55:35 crc kubenswrapper[4799]: I1129 04:55:35.829069 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96493a78-8bf7-49d4-85a0-748217f66262-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "96493a78-8bf7-49d4-85a0-748217f66262" (UID: "96493a78-8bf7-49d4-85a0-748217f66262"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:55:35 crc kubenswrapper[4799]: I1129 04:55:35.893662 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96493a78-8bf7-49d4-85a0-748217f66262-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:35 crc kubenswrapper[4799]: I1129 04:55:35.893714 4799 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/96493a78-8bf7-49d4-85a0-748217f66262-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:35 crc kubenswrapper[4799]: I1129 04:55:35.893726 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96493a78-8bf7-49d4-85a0-748217f66262-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:35 crc kubenswrapper[4799]: I1129 04:55:35.893740 4799 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/96493a78-8bf7-49d4-85a0-748217f66262-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:35 crc kubenswrapper[4799]: I1129 04:55:35.893751 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96493a78-8bf7-49d4-85a0-748217f66262-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:35 crc kubenswrapper[4799]: I1129 04:55:35.893761 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cr2k8\" (UniqueName: \"kubernetes.io/projected/96493a78-8bf7-49d4-85a0-748217f66262-kube-api-access-cr2k8\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:36 crc kubenswrapper[4799]: I1129 04:55:36.264163 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jtjzp" event={"ID":"96493a78-8bf7-49d4-85a0-748217f66262","Type":"ContainerDied","Data":"69264827b6759fc61402a802e0708fffc7501874a55274da76e1237d6b87164c"} Nov 29 04:55:36 crc kubenswrapper[4799]: I1129 04:55:36.264221 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="69264827b6759fc61402a802e0708fffc7501874a55274da76e1237d6b87164c" Nov 29 04:55:36 crc kubenswrapper[4799]: I1129 04:55:36.264254 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jtjzp" Nov 29 04:55:36 crc kubenswrapper[4799]: I1129 04:55:36.374947 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-757bb6ddd6-9wg88"] Nov 29 04:55:36 crc kubenswrapper[4799]: E1129 04:55:36.375528 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96493a78-8bf7-49d4-85a0-748217f66262" containerName="keystone-bootstrap" Nov 29 04:55:36 crc kubenswrapper[4799]: I1129 04:55:36.375556 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="96493a78-8bf7-49d4-85a0-748217f66262" containerName="keystone-bootstrap" Nov 29 04:55:36 crc kubenswrapper[4799]: I1129 04:55:36.375780 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="96493a78-8bf7-49d4-85a0-748217f66262" containerName="keystone-bootstrap" Nov 29 04:55:36 crc kubenswrapper[4799]: I1129 04:55:36.376722 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-757bb6ddd6-9wg88" Nov 29 04:55:36 crc kubenswrapper[4799]: I1129 04:55:36.379402 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 29 04:55:36 crc kubenswrapper[4799]: I1129 04:55:36.382986 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 29 04:55:36 crc kubenswrapper[4799]: I1129 04:55:36.383249 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 29 04:55:36 crc kubenswrapper[4799]: I1129 04:55:36.383478 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-85vg6" Nov 29 04:55:36 crc kubenswrapper[4799]: I1129 04:55:36.383616 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Nov 29 04:55:36 crc kubenswrapper[4799]: I1129 04:55:36.383815 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Nov 29 04:55:36 crc kubenswrapper[4799]: I1129 04:55:36.393511 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-757bb6ddd6-9wg88"] Nov 29 04:55:36 crc kubenswrapper[4799]: I1129 04:55:36.445108 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2ztd\" (UniqueName: \"kubernetes.io/projected/5d5a9855-0646-45ff-8036-afba2a0d1a41-kube-api-access-s2ztd\") pod \"keystone-757bb6ddd6-9wg88\" (UID: \"5d5a9855-0646-45ff-8036-afba2a0d1a41\") " pod="openstack/keystone-757bb6ddd6-9wg88" Nov 29 04:55:36 crc kubenswrapper[4799]: I1129 04:55:36.445171 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d5a9855-0646-45ff-8036-afba2a0d1a41-combined-ca-bundle\") pod \"keystone-757bb6ddd6-9wg88\" (UID: \"5d5a9855-0646-45ff-8036-afba2a0d1a41\") " pod="openstack/keystone-757bb6ddd6-9wg88" Nov 29 04:55:36 crc kubenswrapper[4799]: I1129 04:55:36.445198 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5d5a9855-0646-45ff-8036-afba2a0d1a41-credential-keys\") pod \"keystone-757bb6ddd6-9wg88\" (UID: \"5d5a9855-0646-45ff-8036-afba2a0d1a41\") " pod="openstack/keystone-757bb6ddd6-9wg88" Nov 29 04:55:36 crc kubenswrapper[4799]: I1129 04:55:36.445228 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d5a9855-0646-45ff-8036-afba2a0d1a41-internal-tls-certs\") pod \"keystone-757bb6ddd6-9wg88\" (UID: \"5d5a9855-0646-45ff-8036-afba2a0d1a41\") " pod="openstack/keystone-757bb6ddd6-9wg88" Nov 29 04:55:36 crc kubenswrapper[4799]: I1129 04:55:36.445254 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d5a9855-0646-45ff-8036-afba2a0d1a41-scripts\") pod \"keystone-757bb6ddd6-9wg88\" (UID: \"5d5a9855-0646-45ff-8036-afba2a0d1a41\") " pod="openstack/keystone-757bb6ddd6-9wg88" Nov 29 04:55:36 crc kubenswrapper[4799]: I1129 04:55:36.445268 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d5a9855-0646-45ff-8036-afba2a0d1a41-public-tls-certs\") pod \"keystone-757bb6ddd6-9wg88\" (UID: \"5d5a9855-0646-45ff-8036-afba2a0d1a41\") " pod="openstack/keystone-757bb6ddd6-9wg88" Nov 29 04:55:36 crc kubenswrapper[4799]: I1129 04:55:36.445325 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5d5a9855-0646-45ff-8036-afba2a0d1a41-fernet-keys\") pod \"keystone-757bb6ddd6-9wg88\" (UID: \"5d5a9855-0646-45ff-8036-afba2a0d1a41\") " pod="openstack/keystone-757bb6ddd6-9wg88" Nov 29 04:55:36 crc kubenswrapper[4799]: I1129 04:55:36.445367 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d5a9855-0646-45ff-8036-afba2a0d1a41-config-data\") pod \"keystone-757bb6ddd6-9wg88\" (UID: \"5d5a9855-0646-45ff-8036-afba2a0d1a41\") " pod="openstack/keystone-757bb6ddd6-9wg88" Nov 29 04:55:36 crc kubenswrapper[4799]: I1129 04:55:36.547037 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d5a9855-0646-45ff-8036-afba2a0d1a41-combined-ca-bundle\") pod \"keystone-757bb6ddd6-9wg88\" (UID: \"5d5a9855-0646-45ff-8036-afba2a0d1a41\") " pod="openstack/keystone-757bb6ddd6-9wg88" Nov 29 04:55:36 crc kubenswrapper[4799]: I1129 04:55:36.547090 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5d5a9855-0646-45ff-8036-afba2a0d1a41-credential-keys\") pod \"keystone-757bb6ddd6-9wg88\" (UID: \"5d5a9855-0646-45ff-8036-afba2a0d1a41\") " pod="openstack/keystone-757bb6ddd6-9wg88" Nov 29 04:55:36 crc kubenswrapper[4799]: I1129 04:55:36.547120 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d5a9855-0646-45ff-8036-afba2a0d1a41-internal-tls-certs\") pod \"keystone-757bb6ddd6-9wg88\" (UID: \"5d5a9855-0646-45ff-8036-afba2a0d1a41\") " pod="openstack/keystone-757bb6ddd6-9wg88" Nov 29 04:55:36 crc kubenswrapper[4799]: I1129 04:55:36.547139 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d5a9855-0646-45ff-8036-afba2a0d1a41-scripts\") pod \"keystone-757bb6ddd6-9wg88\" (UID: \"5d5a9855-0646-45ff-8036-afba2a0d1a41\") " pod="openstack/keystone-757bb6ddd6-9wg88" Nov 29 04:55:36 crc kubenswrapper[4799]: I1129 04:55:36.547156 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d5a9855-0646-45ff-8036-afba2a0d1a41-public-tls-certs\") pod \"keystone-757bb6ddd6-9wg88\" (UID: \"5d5a9855-0646-45ff-8036-afba2a0d1a41\") " pod="openstack/keystone-757bb6ddd6-9wg88" Nov 29 04:55:36 crc kubenswrapper[4799]: I1129 04:55:36.547213 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5d5a9855-0646-45ff-8036-afba2a0d1a41-fernet-keys\") pod \"keystone-757bb6ddd6-9wg88\" (UID: \"5d5a9855-0646-45ff-8036-afba2a0d1a41\") " pod="openstack/keystone-757bb6ddd6-9wg88" Nov 29 04:55:36 crc kubenswrapper[4799]: I1129 04:55:36.547253 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d5a9855-0646-45ff-8036-afba2a0d1a41-config-data\") pod \"keystone-757bb6ddd6-9wg88\" (UID: \"5d5a9855-0646-45ff-8036-afba2a0d1a41\") " pod="openstack/keystone-757bb6ddd6-9wg88" Nov 29 04:55:36 crc kubenswrapper[4799]: I1129 04:55:36.547291 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2ztd\" (UniqueName: \"kubernetes.io/projected/5d5a9855-0646-45ff-8036-afba2a0d1a41-kube-api-access-s2ztd\") pod \"keystone-757bb6ddd6-9wg88\" (UID: \"5d5a9855-0646-45ff-8036-afba2a0d1a41\") " pod="openstack/keystone-757bb6ddd6-9wg88" Nov 29 04:55:36 crc kubenswrapper[4799]: I1129 04:55:36.554822 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5d5a9855-0646-45ff-8036-afba2a0d1a41-fernet-keys\") pod \"keystone-757bb6ddd6-9wg88\" (UID: \"5d5a9855-0646-45ff-8036-afba2a0d1a41\") " pod="openstack/keystone-757bb6ddd6-9wg88" Nov 29 04:55:36 crc kubenswrapper[4799]: I1129 04:55:36.555831 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d5a9855-0646-45ff-8036-afba2a0d1a41-scripts\") pod \"keystone-757bb6ddd6-9wg88\" (UID: \"5d5a9855-0646-45ff-8036-afba2a0d1a41\") " pod="openstack/keystone-757bb6ddd6-9wg88" Nov 29 04:55:36 crc kubenswrapper[4799]: I1129 04:55:36.560558 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d5a9855-0646-45ff-8036-afba2a0d1a41-public-tls-certs\") pod \"keystone-757bb6ddd6-9wg88\" (UID: \"5d5a9855-0646-45ff-8036-afba2a0d1a41\") " pod="openstack/keystone-757bb6ddd6-9wg88" Nov 29 04:55:36 crc kubenswrapper[4799]: I1129 04:55:36.561010 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d5a9855-0646-45ff-8036-afba2a0d1a41-combined-ca-bundle\") pod \"keystone-757bb6ddd6-9wg88\" (UID: \"5d5a9855-0646-45ff-8036-afba2a0d1a41\") " pod="openstack/keystone-757bb6ddd6-9wg88" Nov 29 04:55:36 crc kubenswrapper[4799]: I1129 04:55:36.561533 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d5a9855-0646-45ff-8036-afba2a0d1a41-internal-tls-certs\") pod \"keystone-757bb6ddd6-9wg88\" (UID: \"5d5a9855-0646-45ff-8036-afba2a0d1a41\") " pod="openstack/keystone-757bb6ddd6-9wg88" Nov 29 04:55:36 crc kubenswrapper[4799]: I1129 04:55:36.566078 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5d5a9855-0646-45ff-8036-afba2a0d1a41-credential-keys\") pod \"keystone-757bb6ddd6-9wg88\" (UID: \"5d5a9855-0646-45ff-8036-afba2a0d1a41\") " pod="openstack/keystone-757bb6ddd6-9wg88" Nov 29 04:55:36 crc kubenswrapper[4799]: I1129 04:55:36.566885 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d5a9855-0646-45ff-8036-afba2a0d1a41-config-data\") pod \"keystone-757bb6ddd6-9wg88\" (UID: \"5d5a9855-0646-45ff-8036-afba2a0d1a41\") " pod="openstack/keystone-757bb6ddd6-9wg88" Nov 29 04:55:36 crc kubenswrapper[4799]: I1129 04:55:36.568671 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2ztd\" (UniqueName: \"kubernetes.io/projected/5d5a9855-0646-45ff-8036-afba2a0d1a41-kube-api-access-s2ztd\") pod \"keystone-757bb6ddd6-9wg88\" (UID: \"5d5a9855-0646-45ff-8036-afba2a0d1a41\") " pod="openstack/keystone-757bb6ddd6-9wg88" Nov 29 04:55:36 crc kubenswrapper[4799]: I1129 04:55:36.757222 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-757bb6ddd6-9wg88" Nov 29 04:55:38 crc kubenswrapper[4799]: I1129 04:55:38.288300 4799 generic.go:334] "Generic (PLEG): container finished" podID="bcf0c49a-5656-4399-837c-152c9af16309" containerID="cdcd2d082a36adb9d24d954ba9b89729c3c908a3e760003812816af54bcbb49c" exitCode=0 Nov 29 04:55:38 crc kubenswrapper[4799]: I1129 04:55:38.288436 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-9grld" event={"ID":"bcf0c49a-5656-4399-837c-152c9af16309","Type":"ContainerDied","Data":"cdcd2d082a36adb9d24d954ba9b89729c3c908a3e760003812816af54bcbb49c"} Nov 29 04:55:38 crc kubenswrapper[4799]: I1129 04:55:38.522912 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-dc694ffdd-sm2s5"] Nov 29 04:55:38 crc kubenswrapper[4799]: I1129 04:55:38.571237 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-757bb6ddd6-9wg88"] Nov 29 04:55:39 crc kubenswrapper[4799]: I1129 04:55:39.303173 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-757bb6ddd6-9wg88" event={"ID":"5d5a9855-0646-45ff-8036-afba2a0d1a41","Type":"ContainerStarted","Data":"1a2fd3eb2a1f7a472ed81386e9150b17d213a06d6cc708987c052487dc07b1af"} Nov 29 04:55:39 crc kubenswrapper[4799]: I1129 04:55:39.303753 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-757bb6ddd6-9wg88" Nov 29 04:55:39 crc kubenswrapper[4799]: I1129 04:55:39.303776 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-757bb6ddd6-9wg88" event={"ID":"5d5a9855-0646-45ff-8036-afba2a0d1a41","Type":"ContainerStarted","Data":"ef7593de7f4119b93af5b46491783bc5507d818cb2f65de89935fc2ff10e7205"} Nov 29 04:55:39 crc kubenswrapper[4799]: I1129 04:55:39.310740 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"82c42edf-7b25-411f-b00f-95ed3e74bede","Type":"ContainerStarted","Data":"e5541ed59a78cb4ea523c9cbac248a33c6b6584def87b5d8ecd5bac9f2095cd0"} Nov 29 04:55:39 crc kubenswrapper[4799]: I1129 04:55:39.313769 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-dc694ffdd-sm2s5" event={"ID":"8af48d85-03bf-4d5f-a677-49bc96f0dfee","Type":"ContainerStarted","Data":"77f53efeb1ddb1a379b1a72ad6e611278e90d5320b98250aa8c6845986b9e535"} Nov 29 04:55:39 crc kubenswrapper[4799]: I1129 04:55:39.313842 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-dc694ffdd-sm2s5" event={"ID":"8af48d85-03bf-4d5f-a677-49bc96f0dfee","Type":"ContainerStarted","Data":"5da7deff134811f3f1155f207e85dccbb90b51eff7c4020d5b72ca837762bb8c"} Nov 29 04:55:39 crc kubenswrapper[4799]: I1129 04:55:39.313856 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-dc694ffdd-sm2s5" event={"ID":"8af48d85-03bf-4d5f-a677-49bc96f0dfee","Type":"ContainerStarted","Data":"11a0ab1fc065cb02c085e05e224d7adee21c2998873a282776f758ddaea06200"} Nov 29 04:55:39 crc kubenswrapper[4799]: I1129 04:55:39.314128 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-dc694ffdd-sm2s5" Nov 29 04:55:39 crc kubenswrapper[4799]: I1129 04:55:39.314155 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-dc694ffdd-sm2s5" Nov 29 04:55:39 crc kubenswrapper[4799]: I1129 04:55:39.338177 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-757bb6ddd6-9wg88" podStartSLOduration=3.338150084 podStartE2EDuration="3.338150084s" podCreationTimestamp="2025-11-29 04:55:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:55:39.334150316 +0000 UTC m=+1014.977080726" watchObservedRunningTime="2025-11-29 04:55:39.338150084 +0000 UTC m=+1014.981080494" Nov 29 04:55:39 crc kubenswrapper[4799]: I1129 04:55:39.371811 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-dc694ffdd-sm2s5" podStartSLOduration=5.371751844 podStartE2EDuration="5.371751844s" podCreationTimestamp="2025-11-29 04:55:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:55:39.360355026 +0000 UTC m=+1015.003285426" watchObservedRunningTime="2025-11-29 04:55:39.371751844 +0000 UTC m=+1015.014682244" Nov 29 04:55:39 crc kubenswrapper[4799]: I1129 04:55:39.634302 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-9grld" Nov 29 04:55:39 crc kubenswrapper[4799]: I1129 04:55:39.814279 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/bcf0c49a-5656-4399-837c-152c9af16309-config\") pod \"bcf0c49a-5656-4399-837c-152c9af16309\" (UID: \"bcf0c49a-5656-4399-837c-152c9af16309\") " Nov 29 04:55:39 crc kubenswrapper[4799]: I1129 04:55:39.814959 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcf0c49a-5656-4399-837c-152c9af16309-combined-ca-bundle\") pod \"bcf0c49a-5656-4399-837c-152c9af16309\" (UID: \"bcf0c49a-5656-4399-837c-152c9af16309\") " Nov 29 04:55:39 crc kubenswrapper[4799]: I1129 04:55:39.814999 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2n72n\" (UniqueName: \"kubernetes.io/projected/bcf0c49a-5656-4399-837c-152c9af16309-kube-api-access-2n72n\") pod \"bcf0c49a-5656-4399-837c-152c9af16309\" (UID: \"bcf0c49a-5656-4399-837c-152c9af16309\") " Nov 29 04:55:39 crc kubenswrapper[4799]: I1129 04:55:39.830055 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcf0c49a-5656-4399-837c-152c9af16309-kube-api-access-2n72n" (OuterVolumeSpecName: "kube-api-access-2n72n") pod "bcf0c49a-5656-4399-837c-152c9af16309" (UID: "bcf0c49a-5656-4399-837c-152c9af16309"). InnerVolumeSpecName "kube-api-access-2n72n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:55:39 crc kubenswrapper[4799]: I1129 04:55:39.869213 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcf0c49a-5656-4399-837c-152c9af16309-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bcf0c49a-5656-4399-837c-152c9af16309" (UID: "bcf0c49a-5656-4399-837c-152c9af16309"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:55:39 crc kubenswrapper[4799]: I1129 04:55:39.869299 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcf0c49a-5656-4399-837c-152c9af16309-config" (OuterVolumeSpecName: "config") pod "bcf0c49a-5656-4399-837c-152c9af16309" (UID: "bcf0c49a-5656-4399-837c-152c9af16309"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:55:39 crc kubenswrapper[4799]: I1129 04:55:39.918262 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/bcf0c49a-5656-4399-837c-152c9af16309-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:39 crc kubenswrapper[4799]: I1129 04:55:39.918325 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcf0c49a-5656-4399-837c-152c9af16309-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:39 crc kubenswrapper[4799]: I1129 04:55:39.918350 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2n72n\" (UniqueName: \"kubernetes.io/projected/bcf0c49a-5656-4399-837c-152c9af16309-kube-api-access-2n72n\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:40 crc kubenswrapper[4799]: I1129 04:55:40.327667 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-9grld" event={"ID":"bcf0c49a-5656-4399-837c-152c9af16309","Type":"ContainerDied","Data":"9403fb2aec9ba9689eb737b70975060b393fe7471cb57ab3f502c4abbb198fc2"} Nov 29 04:55:40 crc kubenswrapper[4799]: I1129 04:55:40.327746 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9403fb2aec9ba9689eb737b70975060b393fe7471cb57ab3f502c4abbb198fc2" Nov 29 04:55:40 crc kubenswrapper[4799]: I1129 04:55:40.327879 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-9grld" Nov 29 04:55:40 crc kubenswrapper[4799]: I1129 04:55:40.595603 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7b946d459c-5twn6"] Nov 29 04:55:40 crc kubenswrapper[4799]: E1129 04:55:40.596065 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcf0c49a-5656-4399-837c-152c9af16309" containerName="neutron-db-sync" Nov 29 04:55:40 crc kubenswrapper[4799]: I1129 04:55:40.596082 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcf0c49a-5656-4399-837c-152c9af16309" containerName="neutron-db-sync" Nov 29 04:55:40 crc kubenswrapper[4799]: I1129 04:55:40.596293 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcf0c49a-5656-4399-837c-152c9af16309" containerName="neutron-db-sync" Nov 29 04:55:40 crc kubenswrapper[4799]: I1129 04:55:40.597275 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b946d459c-5twn6" Nov 29 04:55:40 crc kubenswrapper[4799]: I1129 04:55:40.607482 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b946d459c-5twn6"] Nov 29 04:55:40 crc kubenswrapper[4799]: I1129 04:55:40.634072 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4-ovsdbserver-nb\") pod \"dnsmasq-dns-7b946d459c-5twn6\" (UID: \"5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4\") " pod="openstack/dnsmasq-dns-7b946d459c-5twn6" Nov 29 04:55:40 crc kubenswrapper[4799]: I1129 04:55:40.634145 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6k27\" (UniqueName: \"kubernetes.io/projected/5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4-kube-api-access-l6k27\") pod \"dnsmasq-dns-7b946d459c-5twn6\" (UID: \"5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4\") " pod="openstack/dnsmasq-dns-7b946d459c-5twn6" Nov 29 04:55:40 crc kubenswrapper[4799]: I1129 04:55:40.634181 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4-ovsdbserver-sb\") pod \"dnsmasq-dns-7b946d459c-5twn6\" (UID: \"5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4\") " pod="openstack/dnsmasq-dns-7b946d459c-5twn6" Nov 29 04:55:40 crc kubenswrapper[4799]: I1129 04:55:40.634489 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4-config\") pod \"dnsmasq-dns-7b946d459c-5twn6\" (UID: \"5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4\") " pod="openstack/dnsmasq-dns-7b946d459c-5twn6" Nov 29 04:55:40 crc kubenswrapper[4799]: I1129 04:55:40.634532 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4-dns-svc\") pod \"dnsmasq-dns-7b946d459c-5twn6\" (UID: \"5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4\") " pod="openstack/dnsmasq-dns-7b946d459c-5twn6" Nov 29 04:55:40 crc kubenswrapper[4799]: I1129 04:55:40.733182 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-d557d86d8-6zk7x"] Nov 29 04:55:40 crc kubenswrapper[4799]: I1129 04:55:40.735123 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d557d86d8-6zk7x" Nov 29 04:55:40 crc kubenswrapper[4799]: I1129 04:55:40.736180 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6k27\" (UniqueName: \"kubernetes.io/projected/5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4-kube-api-access-l6k27\") pod \"dnsmasq-dns-7b946d459c-5twn6\" (UID: \"5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4\") " pod="openstack/dnsmasq-dns-7b946d459c-5twn6" Nov 29 04:55:40 crc kubenswrapper[4799]: I1129 04:55:40.736259 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4-ovsdbserver-sb\") pod \"dnsmasq-dns-7b946d459c-5twn6\" (UID: \"5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4\") " pod="openstack/dnsmasq-dns-7b946d459c-5twn6" Nov 29 04:55:40 crc kubenswrapper[4799]: I1129 04:55:40.736368 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4-config\") pod \"dnsmasq-dns-7b946d459c-5twn6\" (UID: \"5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4\") " pod="openstack/dnsmasq-dns-7b946d459c-5twn6" Nov 29 04:55:40 crc kubenswrapper[4799]: I1129 04:55:40.736394 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4-dns-svc\") pod \"dnsmasq-dns-7b946d459c-5twn6\" (UID: \"5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4\") " pod="openstack/dnsmasq-dns-7b946d459c-5twn6" Nov 29 04:55:40 crc kubenswrapper[4799]: I1129 04:55:40.736465 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4-ovsdbserver-nb\") pod \"dnsmasq-dns-7b946d459c-5twn6\" (UID: \"5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4\") " pod="openstack/dnsmasq-dns-7b946d459c-5twn6" Nov 29 04:55:40 crc kubenswrapper[4799]: I1129 04:55:40.737863 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4-config\") pod \"dnsmasq-dns-7b946d459c-5twn6\" (UID: \"5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4\") " pod="openstack/dnsmasq-dns-7b946d459c-5twn6" Nov 29 04:55:40 crc kubenswrapper[4799]: I1129 04:55:40.737946 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4-ovsdbserver-nb\") pod \"dnsmasq-dns-7b946d459c-5twn6\" (UID: \"5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4\") " pod="openstack/dnsmasq-dns-7b946d459c-5twn6" Nov 29 04:55:40 crc kubenswrapper[4799]: I1129 04:55:40.741205 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4-dns-svc\") pod \"dnsmasq-dns-7b946d459c-5twn6\" (UID: \"5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4\") " pod="openstack/dnsmasq-dns-7b946d459c-5twn6" Nov 29 04:55:40 crc kubenswrapper[4799]: I1129 04:55:40.741993 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4-ovsdbserver-sb\") pod \"dnsmasq-dns-7b946d459c-5twn6\" (UID: \"5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4\") " pod="openstack/dnsmasq-dns-7b946d459c-5twn6" Nov 29 04:55:40 crc kubenswrapper[4799]: I1129 04:55:40.767123 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-d557d86d8-6zk7x"] Nov 29 04:55:40 crc kubenswrapper[4799]: I1129 04:55:40.767576 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-mqltq" Nov 29 04:55:40 crc kubenswrapper[4799]: I1129 04:55:40.768998 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 29 04:55:40 crc kubenswrapper[4799]: I1129 04:55:40.769664 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Nov 29 04:55:40 crc kubenswrapper[4799]: I1129 04:55:40.787120 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 29 04:55:40 crc kubenswrapper[4799]: I1129 04:55:40.828611 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6k27\" (UniqueName: \"kubernetes.io/projected/5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4-kube-api-access-l6k27\") pod \"dnsmasq-dns-7b946d459c-5twn6\" (UID: \"5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4\") " pod="openstack/dnsmasq-dns-7b946d459c-5twn6" Nov 29 04:55:40 crc kubenswrapper[4799]: I1129 04:55:40.837188 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f21cec8f-8cd4-4ecd-9800-3bb0c482fb66-config\") pod \"neutron-d557d86d8-6zk7x\" (UID: \"f21cec8f-8cd4-4ecd-9800-3bb0c482fb66\") " pod="openstack/neutron-d557d86d8-6zk7x" Nov 29 04:55:40 crc kubenswrapper[4799]: I1129 04:55:40.837425 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/f21cec8f-8cd4-4ecd-9800-3bb0c482fb66-httpd-config\") pod \"neutron-d557d86d8-6zk7x\" (UID: \"f21cec8f-8cd4-4ecd-9800-3bb0c482fb66\") " pod="openstack/neutron-d557d86d8-6zk7x" Nov 29 04:55:40 crc kubenswrapper[4799]: I1129 04:55:40.837457 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hm44\" (UniqueName: \"kubernetes.io/projected/f21cec8f-8cd4-4ecd-9800-3bb0c482fb66-kube-api-access-9hm44\") pod \"neutron-d557d86d8-6zk7x\" (UID: \"f21cec8f-8cd4-4ecd-9800-3bb0c482fb66\") " pod="openstack/neutron-d557d86d8-6zk7x" Nov 29 04:55:40 crc kubenswrapper[4799]: I1129 04:55:40.837523 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f21cec8f-8cd4-4ecd-9800-3bb0c482fb66-ovndb-tls-certs\") pod \"neutron-d557d86d8-6zk7x\" (UID: \"f21cec8f-8cd4-4ecd-9800-3bb0c482fb66\") " pod="openstack/neutron-d557d86d8-6zk7x" Nov 29 04:55:40 crc kubenswrapper[4799]: I1129 04:55:40.837605 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f21cec8f-8cd4-4ecd-9800-3bb0c482fb66-combined-ca-bundle\") pod \"neutron-d557d86d8-6zk7x\" (UID: \"f21cec8f-8cd4-4ecd-9800-3bb0c482fb66\") " pod="openstack/neutron-d557d86d8-6zk7x" Nov 29 04:55:40 crc kubenswrapper[4799]: I1129 04:55:40.940147 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f21cec8f-8cd4-4ecd-9800-3bb0c482fb66-config\") pod \"neutron-d557d86d8-6zk7x\" (UID: \"f21cec8f-8cd4-4ecd-9800-3bb0c482fb66\") " pod="openstack/neutron-d557d86d8-6zk7x" Nov 29 04:55:40 crc kubenswrapper[4799]: I1129 04:55:40.940295 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/f21cec8f-8cd4-4ecd-9800-3bb0c482fb66-httpd-config\") pod \"neutron-d557d86d8-6zk7x\" (UID: \"f21cec8f-8cd4-4ecd-9800-3bb0c482fb66\") " pod="openstack/neutron-d557d86d8-6zk7x" Nov 29 04:55:40 crc kubenswrapper[4799]: I1129 04:55:40.940331 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hm44\" (UniqueName: \"kubernetes.io/projected/f21cec8f-8cd4-4ecd-9800-3bb0c482fb66-kube-api-access-9hm44\") pod \"neutron-d557d86d8-6zk7x\" (UID: \"f21cec8f-8cd4-4ecd-9800-3bb0c482fb66\") " pod="openstack/neutron-d557d86d8-6zk7x" Nov 29 04:55:40 crc kubenswrapper[4799]: I1129 04:55:40.940362 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f21cec8f-8cd4-4ecd-9800-3bb0c482fb66-ovndb-tls-certs\") pod \"neutron-d557d86d8-6zk7x\" (UID: \"f21cec8f-8cd4-4ecd-9800-3bb0c482fb66\") " pod="openstack/neutron-d557d86d8-6zk7x" Nov 29 04:55:40 crc kubenswrapper[4799]: I1129 04:55:40.940431 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f21cec8f-8cd4-4ecd-9800-3bb0c482fb66-combined-ca-bundle\") pod \"neutron-d557d86d8-6zk7x\" (UID: \"f21cec8f-8cd4-4ecd-9800-3bb0c482fb66\") " pod="openstack/neutron-d557d86d8-6zk7x" Nov 29 04:55:40 crc kubenswrapper[4799]: I1129 04:55:40.946250 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f21cec8f-8cd4-4ecd-9800-3bb0c482fb66-ovndb-tls-certs\") pod \"neutron-d557d86d8-6zk7x\" (UID: \"f21cec8f-8cd4-4ecd-9800-3bb0c482fb66\") " pod="openstack/neutron-d557d86d8-6zk7x" Nov 29 04:55:40 crc kubenswrapper[4799]: I1129 04:55:40.946509 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f21cec8f-8cd4-4ecd-9800-3bb0c482fb66-combined-ca-bundle\") pod \"neutron-d557d86d8-6zk7x\" (UID: \"f21cec8f-8cd4-4ecd-9800-3bb0c482fb66\") " pod="openstack/neutron-d557d86d8-6zk7x" Nov 29 04:55:40 crc kubenswrapper[4799]: I1129 04:55:40.946946 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/f21cec8f-8cd4-4ecd-9800-3bb0c482fb66-config\") pod \"neutron-d557d86d8-6zk7x\" (UID: \"f21cec8f-8cd4-4ecd-9800-3bb0c482fb66\") " pod="openstack/neutron-d557d86d8-6zk7x" Nov 29 04:55:40 crc kubenswrapper[4799]: I1129 04:55:40.948318 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/f21cec8f-8cd4-4ecd-9800-3bb0c482fb66-httpd-config\") pod \"neutron-d557d86d8-6zk7x\" (UID: \"f21cec8f-8cd4-4ecd-9800-3bb0c482fb66\") " pod="openstack/neutron-d557d86d8-6zk7x" Nov 29 04:55:40 crc kubenswrapper[4799]: I1129 04:55:40.951079 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b946d459c-5twn6" Nov 29 04:55:40 crc kubenswrapper[4799]: I1129 04:55:40.965686 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hm44\" (UniqueName: \"kubernetes.io/projected/f21cec8f-8cd4-4ecd-9800-3bb0c482fb66-kube-api-access-9hm44\") pod \"neutron-d557d86d8-6zk7x\" (UID: \"f21cec8f-8cd4-4ecd-9800-3bb0c482fb66\") " pod="openstack/neutron-d557d86d8-6zk7x" Nov 29 04:55:41 crc kubenswrapper[4799]: I1129 04:55:41.081261 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d557d86d8-6zk7x" Nov 29 04:55:41 crc kubenswrapper[4799]: W1129 04:55:41.465597 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5f0b2f98_90ce_4ca1_84b5_6d88e609a9e4.slice/crio-3d02ebd02fc142f12edc35b6fa85a69ae24d3bf548adcaabb6fd35dc672e92ce WatchSource:0}: Error finding container 3d02ebd02fc142f12edc35b6fa85a69ae24d3bf548adcaabb6fd35dc672e92ce: Status 404 returned error can't find the container with id 3d02ebd02fc142f12edc35b6fa85a69ae24d3bf548adcaabb6fd35dc672e92ce Nov 29 04:55:41 crc kubenswrapper[4799]: I1129 04:55:41.466004 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b946d459c-5twn6"] Nov 29 04:55:41 crc kubenswrapper[4799]: I1129 04:55:41.704939 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-d557d86d8-6zk7x"] Nov 29 04:55:41 crc kubenswrapper[4799]: W1129 04:55:41.774687 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf21cec8f_8cd4_4ecd_9800_3bb0c482fb66.slice/crio-6ff3e3929129c98355795f8e80b1e4d66212c0f3d50ff9159807270bd2648c2d WatchSource:0}: Error finding container 6ff3e3929129c98355795f8e80b1e4d66212c0f3d50ff9159807270bd2648c2d: Status 404 returned error can't find the container with id 6ff3e3929129c98355795f8e80b1e4d66212c0f3d50ff9159807270bd2648c2d Nov 29 04:55:42 crc kubenswrapper[4799]: I1129 04:55:42.355634 4799 generic.go:334] "Generic (PLEG): container finished" podID="5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4" containerID="c1ebde39a09b30f1ca037febc5b83e17a79fd45b252c8afad343b06f831a98ee" exitCode=0 Nov 29 04:55:42 crc kubenswrapper[4799]: I1129 04:55:42.355706 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b946d459c-5twn6" event={"ID":"5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4","Type":"ContainerDied","Data":"c1ebde39a09b30f1ca037febc5b83e17a79fd45b252c8afad343b06f831a98ee"} Nov 29 04:55:42 crc kubenswrapper[4799]: I1129 04:55:42.356063 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b946d459c-5twn6" event={"ID":"5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4","Type":"ContainerStarted","Data":"3d02ebd02fc142f12edc35b6fa85a69ae24d3bf548adcaabb6fd35dc672e92ce"} Nov 29 04:55:42 crc kubenswrapper[4799]: I1129 04:55:42.363638 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d557d86d8-6zk7x" event={"ID":"f21cec8f-8cd4-4ecd-9800-3bb0c482fb66","Type":"ContainerStarted","Data":"0206e122fb25f53bced3ad79d205461cb91bdecdbdf24990b6664f0037824b4c"} Nov 29 04:55:42 crc kubenswrapper[4799]: I1129 04:55:42.363715 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d557d86d8-6zk7x" event={"ID":"f21cec8f-8cd4-4ecd-9800-3bb0c482fb66","Type":"ContainerStarted","Data":"6ff3e3929129c98355795f8e80b1e4d66212c0f3d50ff9159807270bd2648c2d"} Nov 29 04:55:42 crc kubenswrapper[4799]: I1129 04:55:42.373127 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-gcfzn" event={"ID":"8370f287-fa77-4ef5-b929-01a4cf8c598c","Type":"ContainerStarted","Data":"bab0fee75117b65e087b33a56d3dfb3ef3bf37c7c76d8e115b396b47071a0bb7"} Nov 29 04:55:42 crc kubenswrapper[4799]: I1129 04:55:42.413514 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-gcfzn" podStartSLOduration=2.400040519 podStartE2EDuration="45.413493567s" podCreationTimestamp="2025-11-29 04:54:57 +0000 UTC" firstStartedPulling="2025-11-29 04:54:58.382489799 +0000 UTC m=+974.025420199" lastFinishedPulling="2025-11-29 04:55:41.395942847 +0000 UTC m=+1017.038873247" observedRunningTime="2025-11-29 04:55:42.397816625 +0000 UTC m=+1018.040747025" watchObservedRunningTime="2025-11-29 04:55:42.413493567 +0000 UTC m=+1018.056423967" Nov 29 04:55:42 crc kubenswrapper[4799]: I1129 04:55:42.982953 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-669f887b5-74klb"] Nov 29 04:55:42 crc kubenswrapper[4799]: I1129 04:55:42.991156 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-669f887b5-74klb" Nov 29 04:55:42 crc kubenswrapper[4799]: I1129 04:55:42.993469 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Nov 29 04:55:42 crc kubenswrapper[4799]: I1129 04:55:42.993499 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-669f887b5-74klb"] Nov 29 04:55:42 crc kubenswrapper[4799]: I1129 04:55:42.993664 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Nov 29 04:55:43 crc kubenswrapper[4799]: I1129 04:55:43.183292 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vk42q\" (UniqueName: \"kubernetes.io/projected/122be3fa-bc40-45f0-808d-fece3e876c43-kube-api-access-vk42q\") pod \"neutron-669f887b5-74klb\" (UID: \"122be3fa-bc40-45f0-808d-fece3e876c43\") " pod="openstack/neutron-669f887b5-74klb" Nov 29 04:55:43 crc kubenswrapper[4799]: I1129 04:55:43.183419 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/122be3fa-bc40-45f0-808d-fece3e876c43-httpd-config\") pod \"neutron-669f887b5-74klb\" (UID: \"122be3fa-bc40-45f0-808d-fece3e876c43\") " pod="openstack/neutron-669f887b5-74klb" Nov 29 04:55:43 crc kubenswrapper[4799]: I1129 04:55:43.183487 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/122be3fa-bc40-45f0-808d-fece3e876c43-config\") pod \"neutron-669f887b5-74klb\" (UID: \"122be3fa-bc40-45f0-808d-fece3e876c43\") " pod="openstack/neutron-669f887b5-74klb" Nov 29 04:55:43 crc kubenswrapper[4799]: I1129 04:55:43.183530 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/122be3fa-bc40-45f0-808d-fece3e876c43-combined-ca-bundle\") pod \"neutron-669f887b5-74klb\" (UID: \"122be3fa-bc40-45f0-808d-fece3e876c43\") " pod="openstack/neutron-669f887b5-74klb" Nov 29 04:55:43 crc kubenswrapper[4799]: I1129 04:55:43.183561 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/122be3fa-bc40-45f0-808d-fece3e876c43-public-tls-certs\") pod \"neutron-669f887b5-74klb\" (UID: \"122be3fa-bc40-45f0-808d-fece3e876c43\") " pod="openstack/neutron-669f887b5-74klb" Nov 29 04:55:43 crc kubenswrapper[4799]: I1129 04:55:43.183815 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/122be3fa-bc40-45f0-808d-fece3e876c43-ovndb-tls-certs\") pod \"neutron-669f887b5-74klb\" (UID: \"122be3fa-bc40-45f0-808d-fece3e876c43\") " pod="openstack/neutron-669f887b5-74klb" Nov 29 04:55:43 crc kubenswrapper[4799]: I1129 04:55:43.184117 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/122be3fa-bc40-45f0-808d-fece3e876c43-internal-tls-certs\") pod \"neutron-669f887b5-74klb\" (UID: \"122be3fa-bc40-45f0-808d-fece3e876c43\") " pod="openstack/neutron-669f887b5-74klb" Nov 29 04:55:43 crc kubenswrapper[4799]: I1129 04:55:43.285676 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/122be3fa-bc40-45f0-808d-fece3e876c43-ovndb-tls-certs\") pod \"neutron-669f887b5-74klb\" (UID: \"122be3fa-bc40-45f0-808d-fece3e876c43\") " pod="openstack/neutron-669f887b5-74klb" Nov 29 04:55:43 crc kubenswrapper[4799]: I1129 04:55:43.285779 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/122be3fa-bc40-45f0-808d-fece3e876c43-internal-tls-certs\") pod \"neutron-669f887b5-74klb\" (UID: \"122be3fa-bc40-45f0-808d-fece3e876c43\") " pod="openstack/neutron-669f887b5-74klb" Nov 29 04:55:43 crc kubenswrapper[4799]: I1129 04:55:43.285982 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vk42q\" (UniqueName: \"kubernetes.io/projected/122be3fa-bc40-45f0-808d-fece3e876c43-kube-api-access-vk42q\") pod \"neutron-669f887b5-74klb\" (UID: \"122be3fa-bc40-45f0-808d-fece3e876c43\") " pod="openstack/neutron-669f887b5-74klb" Nov 29 04:55:43 crc kubenswrapper[4799]: I1129 04:55:43.286643 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/122be3fa-bc40-45f0-808d-fece3e876c43-httpd-config\") pod \"neutron-669f887b5-74klb\" (UID: \"122be3fa-bc40-45f0-808d-fece3e876c43\") " pod="openstack/neutron-669f887b5-74klb" Nov 29 04:55:43 crc kubenswrapper[4799]: I1129 04:55:43.286751 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/122be3fa-bc40-45f0-808d-fece3e876c43-config\") pod \"neutron-669f887b5-74klb\" (UID: \"122be3fa-bc40-45f0-808d-fece3e876c43\") " pod="openstack/neutron-669f887b5-74klb" Nov 29 04:55:43 crc kubenswrapper[4799]: I1129 04:55:43.287360 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/122be3fa-bc40-45f0-808d-fece3e876c43-combined-ca-bundle\") pod \"neutron-669f887b5-74klb\" (UID: \"122be3fa-bc40-45f0-808d-fece3e876c43\") " pod="openstack/neutron-669f887b5-74klb" Nov 29 04:55:43 crc kubenswrapper[4799]: I1129 04:55:43.287441 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/122be3fa-bc40-45f0-808d-fece3e876c43-public-tls-certs\") pod \"neutron-669f887b5-74klb\" (UID: \"122be3fa-bc40-45f0-808d-fece3e876c43\") " pod="openstack/neutron-669f887b5-74klb" Nov 29 04:55:43 crc kubenswrapper[4799]: I1129 04:55:43.291492 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/122be3fa-bc40-45f0-808d-fece3e876c43-config\") pod \"neutron-669f887b5-74klb\" (UID: \"122be3fa-bc40-45f0-808d-fece3e876c43\") " pod="openstack/neutron-669f887b5-74klb" Nov 29 04:55:43 crc kubenswrapper[4799]: I1129 04:55:43.291629 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/122be3fa-bc40-45f0-808d-fece3e876c43-ovndb-tls-certs\") pod \"neutron-669f887b5-74klb\" (UID: \"122be3fa-bc40-45f0-808d-fece3e876c43\") " pod="openstack/neutron-669f887b5-74klb" Nov 29 04:55:43 crc kubenswrapper[4799]: I1129 04:55:43.292507 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/122be3fa-bc40-45f0-808d-fece3e876c43-internal-tls-certs\") pod \"neutron-669f887b5-74klb\" (UID: \"122be3fa-bc40-45f0-808d-fece3e876c43\") " pod="openstack/neutron-669f887b5-74klb" Nov 29 04:55:43 crc kubenswrapper[4799]: I1129 04:55:43.292567 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/122be3fa-bc40-45f0-808d-fece3e876c43-httpd-config\") pod \"neutron-669f887b5-74klb\" (UID: \"122be3fa-bc40-45f0-808d-fece3e876c43\") " pod="openstack/neutron-669f887b5-74klb" Nov 29 04:55:43 crc kubenswrapper[4799]: I1129 04:55:43.293501 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/122be3fa-bc40-45f0-808d-fece3e876c43-public-tls-certs\") pod \"neutron-669f887b5-74klb\" (UID: \"122be3fa-bc40-45f0-808d-fece3e876c43\") " pod="openstack/neutron-669f887b5-74klb" Nov 29 04:55:43 crc kubenswrapper[4799]: I1129 04:55:43.297331 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/122be3fa-bc40-45f0-808d-fece3e876c43-combined-ca-bundle\") pod \"neutron-669f887b5-74klb\" (UID: \"122be3fa-bc40-45f0-808d-fece3e876c43\") " pod="openstack/neutron-669f887b5-74klb" Nov 29 04:55:43 crc kubenswrapper[4799]: I1129 04:55:43.308319 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vk42q\" (UniqueName: \"kubernetes.io/projected/122be3fa-bc40-45f0-808d-fece3e876c43-kube-api-access-vk42q\") pod \"neutron-669f887b5-74klb\" (UID: \"122be3fa-bc40-45f0-808d-fece3e876c43\") " pod="openstack/neutron-669f887b5-74klb" Nov 29 04:55:43 crc kubenswrapper[4799]: I1129 04:55:43.366145 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-669f887b5-74klb" Nov 29 04:55:43 crc kubenswrapper[4799]: I1129 04:55:43.385960 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d557d86d8-6zk7x" event={"ID":"f21cec8f-8cd4-4ecd-9800-3bb0c482fb66","Type":"ContainerStarted","Data":"af15be66d22796a789def34857a97e75d27b749e255595a4c9a405ba71d3db05"} Nov 29 04:55:43 crc kubenswrapper[4799]: I1129 04:55:43.386383 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-d557d86d8-6zk7x" Nov 29 04:55:43 crc kubenswrapper[4799]: I1129 04:55:43.388359 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b946d459c-5twn6" event={"ID":"5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4","Type":"ContainerStarted","Data":"f3ed7257ca7688115debcec97e98add603e3a40bc3fdef305cf5a91dd0e256cb"} Nov 29 04:55:43 crc kubenswrapper[4799]: I1129 04:55:43.388531 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7b946d459c-5twn6" Nov 29 04:55:43 crc kubenswrapper[4799]: I1129 04:55:43.414560 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-d557d86d8-6zk7x" podStartSLOduration=3.414531044 podStartE2EDuration="3.414531044s" podCreationTimestamp="2025-11-29 04:55:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:55:43.411474899 +0000 UTC m=+1019.054405309" watchObservedRunningTime="2025-11-29 04:55:43.414531044 +0000 UTC m=+1019.057461444" Nov 29 04:55:43 crc kubenswrapper[4799]: I1129 04:55:43.432012 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7b946d459c-5twn6" podStartSLOduration=3.43198002 podStartE2EDuration="3.43198002s" podCreationTimestamp="2025-11-29 04:55:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:55:43.429856468 +0000 UTC m=+1019.072786868" watchObservedRunningTime="2025-11-29 04:55:43.43198002 +0000 UTC m=+1019.074910420" Nov 29 04:55:44 crc kubenswrapper[4799]: I1129 04:55:44.406615 4799 generic.go:334] "Generic (PLEG): container finished" podID="8370f287-fa77-4ef5-b929-01a4cf8c598c" containerID="bab0fee75117b65e087b33a56d3dfb3ef3bf37c7c76d8e115b396b47071a0bb7" exitCode=0 Nov 29 04:55:44 crc kubenswrapper[4799]: I1129 04:55:44.406903 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-gcfzn" event={"ID":"8370f287-fa77-4ef5-b929-01a4cf8c598c","Type":"ContainerDied","Data":"bab0fee75117b65e087b33a56d3dfb3ef3bf37c7c76d8e115b396b47071a0bb7"} Nov 29 04:55:46 crc kubenswrapper[4799]: I1129 04:55:46.582039 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-gcfzn" Nov 29 04:55:46 crc kubenswrapper[4799]: I1129 04:55:46.683583 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8370f287-fa77-4ef5-b929-01a4cf8c598c-combined-ca-bundle\") pod \"8370f287-fa77-4ef5-b929-01a4cf8c598c\" (UID: \"8370f287-fa77-4ef5-b929-01a4cf8c598c\") " Nov 29 04:55:46 crc kubenswrapper[4799]: I1129 04:55:46.684232 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8370f287-fa77-4ef5-b929-01a4cf8c598c-db-sync-config-data\") pod \"8370f287-fa77-4ef5-b929-01a4cf8c598c\" (UID: \"8370f287-fa77-4ef5-b929-01a4cf8c598c\") " Nov 29 04:55:46 crc kubenswrapper[4799]: I1129 04:55:46.684413 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mgq57\" (UniqueName: \"kubernetes.io/projected/8370f287-fa77-4ef5-b929-01a4cf8c598c-kube-api-access-mgq57\") pod \"8370f287-fa77-4ef5-b929-01a4cf8c598c\" (UID: \"8370f287-fa77-4ef5-b929-01a4cf8c598c\") " Nov 29 04:55:46 crc kubenswrapper[4799]: I1129 04:55:46.692611 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8370f287-fa77-4ef5-b929-01a4cf8c598c-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "8370f287-fa77-4ef5-b929-01a4cf8c598c" (UID: "8370f287-fa77-4ef5-b929-01a4cf8c598c"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:55:46 crc kubenswrapper[4799]: I1129 04:55:46.693069 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8370f287-fa77-4ef5-b929-01a4cf8c598c-kube-api-access-mgq57" (OuterVolumeSpecName: "kube-api-access-mgq57") pod "8370f287-fa77-4ef5-b929-01a4cf8c598c" (UID: "8370f287-fa77-4ef5-b929-01a4cf8c598c"). InnerVolumeSpecName "kube-api-access-mgq57". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:55:46 crc kubenswrapper[4799]: I1129 04:55:46.713527 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8370f287-fa77-4ef5-b929-01a4cf8c598c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8370f287-fa77-4ef5-b929-01a4cf8c598c" (UID: "8370f287-fa77-4ef5-b929-01a4cf8c598c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:55:46 crc kubenswrapper[4799]: I1129 04:55:46.787680 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mgq57\" (UniqueName: \"kubernetes.io/projected/8370f287-fa77-4ef5-b929-01a4cf8c598c-kube-api-access-mgq57\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:46 crc kubenswrapper[4799]: I1129 04:55:46.787745 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8370f287-fa77-4ef5-b929-01a4cf8c598c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:46 crc kubenswrapper[4799]: I1129 04:55:46.787886 4799 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8370f287-fa77-4ef5-b929-01a4cf8c598c-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:47 crc kubenswrapper[4799]: I1129 04:55:47.444249 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-gcfzn" event={"ID":"8370f287-fa77-4ef5-b929-01a4cf8c598c","Type":"ContainerDied","Data":"6b706758924ab4618bd3d2d352c81159df1871030ddea09a02a3f8a7e722ffd4"} Nov 29 04:55:47 crc kubenswrapper[4799]: I1129 04:55:47.444688 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6b706758924ab4618bd3d2d352c81159df1871030ddea09a02a3f8a7e722ffd4" Nov 29 04:55:47 crc kubenswrapper[4799]: I1129 04:55:47.444434 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-gcfzn" Nov 29 04:55:47 crc kubenswrapper[4799]: I1129 04:55:47.895012 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-6495665bf6-xn8f7"] Nov 29 04:55:47 crc kubenswrapper[4799]: E1129 04:55:47.895910 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8370f287-fa77-4ef5-b929-01a4cf8c598c" containerName="barbican-db-sync" Nov 29 04:55:47 crc kubenswrapper[4799]: I1129 04:55:47.895929 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="8370f287-fa77-4ef5-b929-01a4cf8c598c" containerName="barbican-db-sync" Nov 29 04:55:47 crc kubenswrapper[4799]: I1129 04:55:47.896187 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="8370f287-fa77-4ef5-b929-01a4cf8c598c" containerName="barbican-db-sync" Nov 29 04:55:47 crc kubenswrapper[4799]: I1129 04:55:47.897452 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6495665bf6-xn8f7" Nov 29 04:55:47 crc kubenswrapper[4799]: I1129 04:55:47.902359 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-dhn66" Nov 29 04:55:47 crc kubenswrapper[4799]: I1129 04:55:47.902608 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 29 04:55:47 crc kubenswrapper[4799]: I1129 04:55:47.904146 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Nov 29 04:55:47 crc kubenswrapper[4799]: I1129 04:55:47.912264 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-76859cbf9f-kdxtl"] Nov 29 04:55:47 crc kubenswrapper[4799]: I1129 04:55:47.914203 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-76859cbf9f-kdxtl" Nov 29 04:55:47 crc kubenswrapper[4799]: I1129 04:55:47.918367 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bdd214a8-9094-42ef-b2c6-2452b59596f7-logs\") pod \"barbican-keystone-listener-6495665bf6-xn8f7\" (UID: \"bdd214a8-9094-42ef-b2c6-2452b59596f7\") " pod="openstack/barbican-keystone-listener-6495665bf6-xn8f7" Nov 29 04:55:47 crc kubenswrapper[4799]: I1129 04:55:47.918426 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdd214a8-9094-42ef-b2c6-2452b59596f7-config-data\") pod \"barbican-keystone-listener-6495665bf6-xn8f7\" (UID: \"bdd214a8-9094-42ef-b2c6-2452b59596f7\") " pod="openstack/barbican-keystone-listener-6495665bf6-xn8f7" Nov 29 04:55:47 crc kubenswrapper[4799]: I1129 04:55:47.918496 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bdd214a8-9094-42ef-b2c6-2452b59596f7-config-data-custom\") pod \"barbican-keystone-listener-6495665bf6-xn8f7\" (UID: \"bdd214a8-9094-42ef-b2c6-2452b59596f7\") " pod="openstack/barbican-keystone-listener-6495665bf6-xn8f7" Nov 29 04:55:47 crc kubenswrapper[4799]: I1129 04:55:47.918568 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdpw9\" (UniqueName: \"kubernetes.io/projected/bdd214a8-9094-42ef-b2c6-2452b59596f7-kube-api-access-hdpw9\") pod \"barbican-keystone-listener-6495665bf6-xn8f7\" (UID: \"bdd214a8-9094-42ef-b2c6-2452b59596f7\") " pod="openstack/barbican-keystone-listener-6495665bf6-xn8f7" Nov 29 04:55:47 crc kubenswrapper[4799]: I1129 04:55:47.918953 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdd214a8-9094-42ef-b2c6-2452b59596f7-combined-ca-bundle\") pod \"barbican-keystone-listener-6495665bf6-xn8f7\" (UID: \"bdd214a8-9094-42ef-b2c6-2452b59596f7\") " pod="openstack/barbican-keystone-listener-6495665bf6-xn8f7" Nov 29 04:55:47 crc kubenswrapper[4799]: I1129 04:55:47.925809 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Nov 29 04:55:47 crc kubenswrapper[4799]: I1129 04:55:47.929506 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6495665bf6-xn8f7"] Nov 29 04:55:47 crc kubenswrapper[4799]: I1129 04:55:47.940329 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-76859cbf9f-kdxtl"] Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.020805 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdd214a8-9094-42ef-b2c6-2452b59596f7-config-data\") pod \"barbican-keystone-listener-6495665bf6-xn8f7\" (UID: \"bdd214a8-9094-42ef-b2c6-2452b59596f7\") " pod="openstack/barbican-keystone-listener-6495665bf6-xn8f7" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.020877 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bdd214a8-9094-42ef-b2c6-2452b59596f7-logs\") pod \"barbican-keystone-listener-6495665bf6-xn8f7\" (UID: \"bdd214a8-9094-42ef-b2c6-2452b59596f7\") " pod="openstack/barbican-keystone-listener-6495665bf6-xn8f7" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.020908 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f14d7038-795e-4f8a-8ce7-81e64c86137b-logs\") pod \"barbican-worker-76859cbf9f-kdxtl\" (UID: \"f14d7038-795e-4f8a-8ce7-81e64c86137b\") " pod="openstack/barbican-worker-76859cbf9f-kdxtl" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.020928 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bdd214a8-9094-42ef-b2c6-2452b59596f7-config-data-custom\") pod \"barbican-keystone-listener-6495665bf6-xn8f7\" (UID: \"bdd214a8-9094-42ef-b2c6-2452b59596f7\") " pod="openstack/barbican-keystone-listener-6495665bf6-xn8f7" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.020963 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f14d7038-795e-4f8a-8ce7-81e64c86137b-config-data-custom\") pod \"barbican-worker-76859cbf9f-kdxtl\" (UID: \"f14d7038-795e-4f8a-8ce7-81e64c86137b\") " pod="openstack/barbican-worker-76859cbf9f-kdxtl" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.020986 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f14d7038-795e-4f8a-8ce7-81e64c86137b-combined-ca-bundle\") pod \"barbican-worker-76859cbf9f-kdxtl\" (UID: \"f14d7038-795e-4f8a-8ce7-81e64c86137b\") " pod="openstack/barbican-worker-76859cbf9f-kdxtl" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.021002 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f14d7038-795e-4f8a-8ce7-81e64c86137b-config-data\") pod \"barbican-worker-76859cbf9f-kdxtl\" (UID: \"f14d7038-795e-4f8a-8ce7-81e64c86137b\") " pod="openstack/barbican-worker-76859cbf9f-kdxtl" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.021032 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdpw9\" (UniqueName: \"kubernetes.io/projected/bdd214a8-9094-42ef-b2c6-2452b59596f7-kube-api-access-hdpw9\") pod \"barbican-keystone-listener-6495665bf6-xn8f7\" (UID: \"bdd214a8-9094-42ef-b2c6-2452b59596f7\") " pod="openstack/barbican-keystone-listener-6495665bf6-xn8f7" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.021090 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttrz8\" (UniqueName: \"kubernetes.io/projected/f14d7038-795e-4f8a-8ce7-81e64c86137b-kube-api-access-ttrz8\") pod \"barbican-worker-76859cbf9f-kdxtl\" (UID: \"f14d7038-795e-4f8a-8ce7-81e64c86137b\") " pod="openstack/barbican-worker-76859cbf9f-kdxtl" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.021113 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdd214a8-9094-42ef-b2c6-2452b59596f7-combined-ca-bundle\") pod \"barbican-keystone-listener-6495665bf6-xn8f7\" (UID: \"bdd214a8-9094-42ef-b2c6-2452b59596f7\") " pod="openstack/barbican-keystone-listener-6495665bf6-xn8f7" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.022028 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bdd214a8-9094-42ef-b2c6-2452b59596f7-logs\") pod \"barbican-keystone-listener-6495665bf6-xn8f7\" (UID: \"bdd214a8-9094-42ef-b2c6-2452b59596f7\") " pod="openstack/barbican-keystone-listener-6495665bf6-xn8f7" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.030418 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdd214a8-9094-42ef-b2c6-2452b59596f7-combined-ca-bundle\") pod \"barbican-keystone-listener-6495665bf6-xn8f7\" (UID: \"bdd214a8-9094-42ef-b2c6-2452b59596f7\") " pod="openstack/barbican-keystone-listener-6495665bf6-xn8f7" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.031874 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdd214a8-9094-42ef-b2c6-2452b59596f7-config-data\") pod \"barbican-keystone-listener-6495665bf6-xn8f7\" (UID: \"bdd214a8-9094-42ef-b2c6-2452b59596f7\") " pod="openstack/barbican-keystone-listener-6495665bf6-xn8f7" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.055019 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bdd214a8-9094-42ef-b2c6-2452b59596f7-config-data-custom\") pod \"barbican-keystone-listener-6495665bf6-xn8f7\" (UID: \"bdd214a8-9094-42ef-b2c6-2452b59596f7\") " pod="openstack/barbican-keystone-listener-6495665bf6-xn8f7" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.060428 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdpw9\" (UniqueName: \"kubernetes.io/projected/bdd214a8-9094-42ef-b2c6-2452b59596f7-kube-api-access-hdpw9\") pod \"barbican-keystone-listener-6495665bf6-xn8f7\" (UID: \"bdd214a8-9094-42ef-b2c6-2452b59596f7\") " pod="openstack/barbican-keystone-listener-6495665bf6-xn8f7" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.066248 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b946d459c-5twn6"] Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.067415 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7b946d459c-5twn6" podUID="5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4" containerName="dnsmasq-dns" containerID="cri-o://f3ed7257ca7688115debcec97e98add603e3a40bc3fdef305cf5a91dd0e256cb" gracePeriod=10 Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.085024 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7b946d459c-5twn6" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.121046 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bb684768f-x97jl"] Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.122512 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f14d7038-795e-4f8a-8ce7-81e64c86137b-logs\") pod \"barbican-worker-76859cbf9f-kdxtl\" (UID: \"f14d7038-795e-4f8a-8ce7-81e64c86137b\") " pod="openstack/barbican-worker-76859cbf9f-kdxtl" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.122653 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f14d7038-795e-4f8a-8ce7-81e64c86137b-config-data-custom\") pod \"barbican-worker-76859cbf9f-kdxtl\" (UID: \"f14d7038-795e-4f8a-8ce7-81e64c86137b\") " pod="openstack/barbican-worker-76859cbf9f-kdxtl" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.122758 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f14d7038-795e-4f8a-8ce7-81e64c86137b-combined-ca-bundle\") pod \"barbican-worker-76859cbf9f-kdxtl\" (UID: \"f14d7038-795e-4f8a-8ce7-81e64c86137b\") " pod="openstack/barbican-worker-76859cbf9f-kdxtl" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.122887 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f14d7038-795e-4f8a-8ce7-81e64c86137b-config-data\") pod \"barbican-worker-76859cbf9f-kdxtl\" (UID: \"f14d7038-795e-4f8a-8ce7-81e64c86137b\") " pod="openstack/barbican-worker-76859cbf9f-kdxtl" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.123032 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttrz8\" (UniqueName: \"kubernetes.io/projected/f14d7038-795e-4f8a-8ce7-81e64c86137b-kube-api-access-ttrz8\") pod \"barbican-worker-76859cbf9f-kdxtl\" (UID: \"f14d7038-795e-4f8a-8ce7-81e64c86137b\") " pod="openstack/barbican-worker-76859cbf9f-kdxtl" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.124029 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f14d7038-795e-4f8a-8ce7-81e64c86137b-logs\") pod \"barbican-worker-76859cbf9f-kdxtl\" (UID: \"f14d7038-795e-4f8a-8ce7-81e64c86137b\") " pod="openstack/barbican-worker-76859cbf9f-kdxtl" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.130136 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb684768f-x97jl" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.131599 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f14d7038-795e-4f8a-8ce7-81e64c86137b-combined-ca-bundle\") pod \"barbican-worker-76859cbf9f-kdxtl\" (UID: \"f14d7038-795e-4f8a-8ce7-81e64c86137b\") " pod="openstack/barbican-worker-76859cbf9f-kdxtl" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.138212 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f14d7038-795e-4f8a-8ce7-81e64c86137b-config-data\") pod \"barbican-worker-76859cbf9f-kdxtl\" (UID: \"f14d7038-795e-4f8a-8ce7-81e64c86137b\") " pod="openstack/barbican-worker-76859cbf9f-kdxtl" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.140406 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f14d7038-795e-4f8a-8ce7-81e64c86137b-config-data-custom\") pod \"barbican-worker-76859cbf9f-kdxtl\" (UID: \"f14d7038-795e-4f8a-8ce7-81e64c86137b\") " pod="openstack/barbican-worker-76859cbf9f-kdxtl" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.147814 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttrz8\" (UniqueName: \"kubernetes.io/projected/f14d7038-795e-4f8a-8ce7-81e64c86137b-kube-api-access-ttrz8\") pod \"barbican-worker-76859cbf9f-kdxtl\" (UID: \"f14d7038-795e-4f8a-8ce7-81e64c86137b\") " pod="openstack/barbican-worker-76859cbf9f-kdxtl" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.152862 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bb684768f-x97jl"] Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.224852 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0ae506e6-d401-4fa6-b4ce-8db525ef694e-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb684768f-x97jl\" (UID: \"0ae506e6-d401-4fa6-b4ce-8db525ef694e\") " pod="openstack/dnsmasq-dns-6bb684768f-x97jl" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.224939 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0ae506e6-d401-4fa6-b4ce-8db525ef694e-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb684768f-x97jl\" (UID: \"0ae506e6-d401-4fa6-b4ce-8db525ef694e\") " pod="openstack/dnsmasq-dns-6bb684768f-x97jl" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.224977 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0ae506e6-d401-4fa6-b4ce-8db525ef694e-dns-svc\") pod \"dnsmasq-dns-6bb684768f-x97jl\" (UID: \"0ae506e6-d401-4fa6-b4ce-8db525ef694e\") " pod="openstack/dnsmasq-dns-6bb684768f-x97jl" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.224996 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2f5wf\" (UniqueName: \"kubernetes.io/projected/0ae506e6-d401-4fa6-b4ce-8db525ef694e-kube-api-access-2f5wf\") pod \"dnsmasq-dns-6bb684768f-x97jl\" (UID: \"0ae506e6-d401-4fa6-b4ce-8db525ef694e\") " pod="openstack/dnsmasq-dns-6bb684768f-x97jl" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.225036 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ae506e6-d401-4fa6-b4ce-8db525ef694e-config\") pod \"dnsmasq-dns-6bb684768f-x97jl\" (UID: \"0ae506e6-d401-4fa6-b4ce-8db525ef694e\") " pod="openstack/dnsmasq-dns-6bb684768f-x97jl" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.242777 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6944bdb4b-6bt2h"] Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.244632 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6944bdb4b-6bt2h" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.247662 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6495665bf6-xn8f7" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.248533 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.267564 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6944bdb4b-6bt2h"] Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.281712 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-76859cbf9f-kdxtl" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.326972 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e964f0f8-6843-4a3c-917b-69372833af7a-config-data\") pod \"barbican-api-6944bdb4b-6bt2h\" (UID: \"e964f0f8-6843-4a3c-917b-69372833af7a\") " pod="openstack/barbican-api-6944bdb4b-6bt2h" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.327419 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0ae506e6-d401-4fa6-b4ce-8db525ef694e-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb684768f-x97jl\" (UID: \"0ae506e6-d401-4fa6-b4ce-8db525ef694e\") " pod="openstack/dnsmasq-dns-6bb684768f-x97jl" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.327585 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zngmq\" (UniqueName: \"kubernetes.io/projected/e964f0f8-6843-4a3c-917b-69372833af7a-kube-api-access-zngmq\") pod \"barbican-api-6944bdb4b-6bt2h\" (UID: \"e964f0f8-6843-4a3c-917b-69372833af7a\") " pod="openstack/barbican-api-6944bdb4b-6bt2h" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.327725 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0ae506e6-d401-4fa6-b4ce-8db525ef694e-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb684768f-x97jl\" (UID: \"0ae506e6-d401-4fa6-b4ce-8db525ef694e\") " pod="openstack/dnsmasq-dns-6bb684768f-x97jl" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.328133 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0ae506e6-d401-4fa6-b4ce-8db525ef694e-dns-svc\") pod \"dnsmasq-dns-6bb684768f-x97jl\" (UID: \"0ae506e6-d401-4fa6-b4ce-8db525ef694e\") " pod="openstack/dnsmasq-dns-6bb684768f-x97jl" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.328277 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2f5wf\" (UniqueName: \"kubernetes.io/projected/0ae506e6-d401-4fa6-b4ce-8db525ef694e-kube-api-access-2f5wf\") pod \"dnsmasq-dns-6bb684768f-x97jl\" (UID: \"0ae506e6-d401-4fa6-b4ce-8db525ef694e\") " pod="openstack/dnsmasq-dns-6bb684768f-x97jl" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.328899 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0ae506e6-d401-4fa6-b4ce-8db525ef694e-dns-svc\") pod \"dnsmasq-dns-6bb684768f-x97jl\" (UID: \"0ae506e6-d401-4fa6-b4ce-8db525ef694e\") " pod="openstack/dnsmasq-dns-6bb684768f-x97jl" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.328833 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0ae506e6-d401-4fa6-b4ce-8db525ef694e-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb684768f-x97jl\" (UID: \"0ae506e6-d401-4fa6-b4ce-8db525ef694e\") " pod="openstack/dnsmasq-dns-6bb684768f-x97jl" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.329304 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e964f0f8-6843-4a3c-917b-69372833af7a-combined-ca-bundle\") pod \"barbican-api-6944bdb4b-6bt2h\" (UID: \"e964f0f8-6843-4a3c-917b-69372833af7a\") " pod="openstack/barbican-api-6944bdb4b-6bt2h" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.329337 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0ae506e6-d401-4fa6-b4ce-8db525ef694e-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb684768f-x97jl\" (UID: \"0ae506e6-d401-4fa6-b4ce-8db525ef694e\") " pod="openstack/dnsmasq-dns-6bb684768f-x97jl" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.329439 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ae506e6-d401-4fa6-b4ce-8db525ef694e-config\") pod \"dnsmasq-dns-6bb684768f-x97jl\" (UID: \"0ae506e6-d401-4fa6-b4ce-8db525ef694e\") " pod="openstack/dnsmasq-dns-6bb684768f-x97jl" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.329737 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e964f0f8-6843-4a3c-917b-69372833af7a-config-data-custom\") pod \"barbican-api-6944bdb4b-6bt2h\" (UID: \"e964f0f8-6843-4a3c-917b-69372833af7a\") " pod="openstack/barbican-api-6944bdb4b-6bt2h" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.330040 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e964f0f8-6843-4a3c-917b-69372833af7a-logs\") pod \"barbican-api-6944bdb4b-6bt2h\" (UID: \"e964f0f8-6843-4a3c-917b-69372833af7a\") " pod="openstack/barbican-api-6944bdb4b-6bt2h" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.330224 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ae506e6-d401-4fa6-b4ce-8db525ef694e-config\") pod \"dnsmasq-dns-6bb684768f-x97jl\" (UID: \"0ae506e6-d401-4fa6-b4ce-8db525ef694e\") " pod="openstack/dnsmasq-dns-6bb684768f-x97jl" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.353914 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2f5wf\" (UniqueName: \"kubernetes.io/projected/0ae506e6-d401-4fa6-b4ce-8db525ef694e-kube-api-access-2f5wf\") pod \"dnsmasq-dns-6bb684768f-x97jl\" (UID: \"0ae506e6-d401-4fa6-b4ce-8db525ef694e\") " pod="openstack/dnsmasq-dns-6bb684768f-x97jl" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.432929 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zngmq\" (UniqueName: \"kubernetes.io/projected/e964f0f8-6843-4a3c-917b-69372833af7a-kube-api-access-zngmq\") pod \"barbican-api-6944bdb4b-6bt2h\" (UID: \"e964f0f8-6843-4a3c-917b-69372833af7a\") " pod="openstack/barbican-api-6944bdb4b-6bt2h" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.433055 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e964f0f8-6843-4a3c-917b-69372833af7a-combined-ca-bundle\") pod \"barbican-api-6944bdb4b-6bt2h\" (UID: \"e964f0f8-6843-4a3c-917b-69372833af7a\") " pod="openstack/barbican-api-6944bdb4b-6bt2h" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.433138 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e964f0f8-6843-4a3c-917b-69372833af7a-config-data-custom\") pod \"barbican-api-6944bdb4b-6bt2h\" (UID: \"e964f0f8-6843-4a3c-917b-69372833af7a\") " pod="openstack/barbican-api-6944bdb4b-6bt2h" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.433166 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e964f0f8-6843-4a3c-917b-69372833af7a-logs\") pod \"barbican-api-6944bdb4b-6bt2h\" (UID: \"e964f0f8-6843-4a3c-917b-69372833af7a\") " pod="openstack/barbican-api-6944bdb4b-6bt2h" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.433957 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e964f0f8-6843-4a3c-917b-69372833af7a-config-data\") pod \"barbican-api-6944bdb4b-6bt2h\" (UID: \"e964f0f8-6843-4a3c-917b-69372833af7a\") " pod="openstack/barbican-api-6944bdb4b-6bt2h" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.434727 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e964f0f8-6843-4a3c-917b-69372833af7a-logs\") pod \"barbican-api-6944bdb4b-6bt2h\" (UID: \"e964f0f8-6843-4a3c-917b-69372833af7a\") " pod="openstack/barbican-api-6944bdb4b-6bt2h" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.437679 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e964f0f8-6843-4a3c-917b-69372833af7a-combined-ca-bundle\") pod \"barbican-api-6944bdb4b-6bt2h\" (UID: \"e964f0f8-6843-4a3c-917b-69372833af7a\") " pod="openstack/barbican-api-6944bdb4b-6bt2h" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.438569 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e964f0f8-6843-4a3c-917b-69372833af7a-config-data-custom\") pod \"barbican-api-6944bdb4b-6bt2h\" (UID: \"e964f0f8-6843-4a3c-917b-69372833af7a\") " pod="openstack/barbican-api-6944bdb4b-6bt2h" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.438939 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e964f0f8-6843-4a3c-917b-69372833af7a-config-data\") pod \"barbican-api-6944bdb4b-6bt2h\" (UID: \"e964f0f8-6843-4a3c-917b-69372833af7a\") " pod="openstack/barbican-api-6944bdb4b-6bt2h" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.457120 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zngmq\" (UniqueName: \"kubernetes.io/projected/e964f0f8-6843-4a3c-917b-69372833af7a-kube-api-access-zngmq\") pod \"barbican-api-6944bdb4b-6bt2h\" (UID: \"e964f0f8-6843-4a3c-917b-69372833af7a\") " pod="openstack/barbican-api-6944bdb4b-6bt2h" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.535254 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb684768f-x97jl" Nov 29 04:55:48 crc kubenswrapper[4799]: I1129 04:55:48.576760 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6944bdb4b-6bt2h" Nov 29 04:55:49 crc kubenswrapper[4799]: I1129 04:55:49.444510 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-669f887b5-74klb"] Nov 29 04:55:49 crc kubenswrapper[4799]: I1129 04:55:49.474693 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-669f887b5-74klb" event={"ID":"122be3fa-bc40-45f0-808d-fece3e876c43","Type":"ContainerStarted","Data":"6488680cc28a6a86b49d24cc8d97a4b72d6abdfa17cdfabc1f715240f8b8af73"} Nov 29 04:55:49 crc kubenswrapper[4799]: I1129 04:55:49.488679 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b946d459c-5twn6" Nov 29 04:55:49 crc kubenswrapper[4799]: I1129 04:55:49.489429 4799 generic.go:334] "Generic (PLEG): container finished" podID="5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4" containerID="f3ed7257ca7688115debcec97e98add603e3a40bc3fdef305cf5a91dd0e256cb" exitCode=0 Nov 29 04:55:49 crc kubenswrapper[4799]: I1129 04:55:49.489477 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b946d459c-5twn6" event={"ID":"5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4","Type":"ContainerDied","Data":"f3ed7257ca7688115debcec97e98add603e3a40bc3fdef305cf5a91dd0e256cb"} Nov 29 04:55:49 crc kubenswrapper[4799]: I1129 04:55:49.489511 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b946d459c-5twn6" event={"ID":"5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4","Type":"ContainerDied","Data":"3d02ebd02fc142f12edc35b6fa85a69ae24d3bf548adcaabb6fd35dc672e92ce"} Nov 29 04:55:49 crc kubenswrapper[4799]: I1129 04:55:49.489536 4799 scope.go:117] "RemoveContainer" containerID="f3ed7257ca7688115debcec97e98add603e3a40bc3fdef305cf5a91dd0e256cb" Nov 29 04:55:49 crc kubenswrapper[4799]: I1129 04:55:49.541769 4799 scope.go:117] "RemoveContainer" containerID="c1ebde39a09b30f1ca037febc5b83e17a79fd45b252c8afad343b06f831a98ee" Nov 29 04:55:49 crc kubenswrapper[4799]: I1129 04:55:49.561244 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4-dns-svc\") pod \"5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4\" (UID: \"5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4\") " Nov 29 04:55:49 crc kubenswrapper[4799]: I1129 04:55:49.561411 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4-ovsdbserver-nb\") pod \"5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4\" (UID: \"5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4\") " Nov 29 04:55:49 crc kubenswrapper[4799]: I1129 04:55:49.561597 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4-ovsdbserver-sb\") pod \"5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4\" (UID: \"5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4\") " Nov 29 04:55:49 crc kubenswrapper[4799]: I1129 04:55:49.561621 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l6k27\" (UniqueName: \"kubernetes.io/projected/5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4-kube-api-access-l6k27\") pod \"5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4\" (UID: \"5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4\") " Nov 29 04:55:49 crc kubenswrapper[4799]: I1129 04:55:49.561670 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4-config\") pod \"5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4\" (UID: \"5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4\") " Nov 29 04:55:49 crc kubenswrapper[4799]: I1129 04:55:49.585146 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4-kube-api-access-l6k27" (OuterVolumeSpecName: "kube-api-access-l6k27") pod "5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4" (UID: "5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4"). InnerVolumeSpecName "kube-api-access-l6k27". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:55:49 crc kubenswrapper[4799]: I1129 04:55:49.619896 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6495665bf6-xn8f7"] Nov 29 04:55:49 crc kubenswrapper[4799]: I1129 04:55:49.632361 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bb684768f-x97jl"] Nov 29 04:55:49 crc kubenswrapper[4799]: I1129 04:55:49.641374 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4" (UID: "5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:55:49 crc kubenswrapper[4799]: I1129 04:55:49.653285 4799 scope.go:117] "RemoveContainer" containerID="f3ed7257ca7688115debcec97e98add603e3a40bc3fdef305cf5a91dd0e256cb" Nov 29 04:55:49 crc kubenswrapper[4799]: E1129 04:55:49.655527 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3ed7257ca7688115debcec97e98add603e3a40bc3fdef305cf5a91dd0e256cb\": container with ID starting with f3ed7257ca7688115debcec97e98add603e3a40bc3fdef305cf5a91dd0e256cb not found: ID does not exist" containerID="f3ed7257ca7688115debcec97e98add603e3a40bc3fdef305cf5a91dd0e256cb" Nov 29 04:55:49 crc kubenswrapper[4799]: I1129 04:55:49.655583 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3ed7257ca7688115debcec97e98add603e3a40bc3fdef305cf5a91dd0e256cb"} err="failed to get container status \"f3ed7257ca7688115debcec97e98add603e3a40bc3fdef305cf5a91dd0e256cb\": rpc error: code = NotFound desc = could not find container \"f3ed7257ca7688115debcec97e98add603e3a40bc3fdef305cf5a91dd0e256cb\": container with ID starting with f3ed7257ca7688115debcec97e98add603e3a40bc3fdef305cf5a91dd0e256cb not found: ID does not exist" Nov 29 04:55:49 crc kubenswrapper[4799]: I1129 04:55:49.655637 4799 scope.go:117] "RemoveContainer" containerID="c1ebde39a09b30f1ca037febc5b83e17a79fd45b252c8afad343b06f831a98ee" Nov 29 04:55:49 crc kubenswrapper[4799]: E1129 04:55:49.656409 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1ebde39a09b30f1ca037febc5b83e17a79fd45b252c8afad343b06f831a98ee\": container with ID starting with c1ebde39a09b30f1ca037febc5b83e17a79fd45b252c8afad343b06f831a98ee not found: ID does not exist" containerID="c1ebde39a09b30f1ca037febc5b83e17a79fd45b252c8afad343b06f831a98ee" Nov 29 04:55:49 crc kubenswrapper[4799]: I1129 04:55:49.656451 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1ebde39a09b30f1ca037febc5b83e17a79fd45b252c8afad343b06f831a98ee"} err="failed to get container status \"c1ebde39a09b30f1ca037febc5b83e17a79fd45b252c8afad343b06f831a98ee\": rpc error: code = NotFound desc = could not find container \"c1ebde39a09b30f1ca037febc5b83e17a79fd45b252c8afad343b06f831a98ee\": container with ID starting with c1ebde39a09b30f1ca037febc5b83e17a79fd45b252c8afad343b06f831a98ee not found: ID does not exist" Nov 29 04:55:49 crc kubenswrapper[4799]: I1129 04:55:49.662949 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4" (UID: "5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:55:49 crc kubenswrapper[4799]: I1129 04:55:49.663367 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4-dns-svc\") pod \"5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4\" (UID: \"5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4\") " Nov 29 04:55:49 crc kubenswrapper[4799]: W1129 04:55:49.663629 4799 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4/volumes/kubernetes.io~configmap/dns-svc Nov 29 04:55:49 crc kubenswrapper[4799]: I1129 04:55:49.663653 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4" (UID: "5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:55:49 crc kubenswrapper[4799]: I1129 04:55:49.664902 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:49 crc kubenswrapper[4799]: I1129 04:55:49.664928 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l6k27\" (UniqueName: \"kubernetes.io/projected/5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4-kube-api-access-l6k27\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:49 crc kubenswrapper[4799]: I1129 04:55:49.664943 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:49 crc kubenswrapper[4799]: I1129 04:55:49.668733 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-76859cbf9f-kdxtl"] Nov 29 04:55:49 crc kubenswrapper[4799]: I1129 04:55:49.678666 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4" (UID: "5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:55:49 crc kubenswrapper[4799]: I1129 04:55:49.694904 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4-config" (OuterVolumeSpecName: "config") pod "5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4" (UID: "5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:55:49 crc kubenswrapper[4799]: I1129 04:55:49.766473 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:49 crc kubenswrapper[4799]: I1129 04:55:49.766510 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:49 crc kubenswrapper[4799]: I1129 04:55:49.776049 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6944bdb4b-6bt2h"] Nov 29 04:55:49 crc kubenswrapper[4799]: W1129 04:55:49.789843 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode964f0f8_6843_4a3c_917b_69372833af7a.slice/crio-c001f01269ba72b4d3b08214fc658088bb0920bdb8a9a18f6ea9c52cf5028e9d WatchSource:0}: Error finding container c001f01269ba72b4d3b08214fc658088bb0920bdb8a9a18f6ea9c52cf5028e9d: Status 404 returned error can't find the container with id c001f01269ba72b4d3b08214fc658088bb0920bdb8a9a18f6ea9c52cf5028e9d Nov 29 04:55:50 crc kubenswrapper[4799]: I1129 04:55:50.522776 4799 generic.go:334] "Generic (PLEG): container finished" podID="0ae506e6-d401-4fa6-b4ce-8db525ef694e" containerID="d9fa614974d8b4dda6d470d2fddb9c45c4a2825e1d075b1fe65142f6749caba5" exitCode=0 Nov 29 04:55:50 crc kubenswrapper[4799]: I1129 04:55:50.523367 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb684768f-x97jl" event={"ID":"0ae506e6-d401-4fa6-b4ce-8db525ef694e","Type":"ContainerDied","Data":"d9fa614974d8b4dda6d470d2fddb9c45c4a2825e1d075b1fe65142f6749caba5"} Nov 29 04:55:50 crc kubenswrapper[4799]: I1129 04:55:50.523406 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb684768f-x97jl" event={"ID":"0ae506e6-d401-4fa6-b4ce-8db525ef694e","Type":"ContainerStarted","Data":"c5f4c51a4869b03f3a4338ee5573abf6799fbc1e6c086a3f1aec8b9426b57642"} Nov 29 04:55:50 crc kubenswrapper[4799]: I1129 04:55:50.530262 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b946d459c-5twn6" Nov 29 04:55:50 crc kubenswrapper[4799]: I1129 04:55:50.534072 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6944bdb4b-6bt2h" event={"ID":"e964f0f8-6843-4a3c-917b-69372833af7a","Type":"ContainerStarted","Data":"f073508318cdb64c0905a420f0d8f81d6ed3345066cadbf338a682fe615b86a9"} Nov 29 04:55:50 crc kubenswrapper[4799]: I1129 04:55:50.534152 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6944bdb4b-6bt2h" event={"ID":"e964f0f8-6843-4a3c-917b-69372833af7a","Type":"ContainerStarted","Data":"c001f01269ba72b4d3b08214fc658088bb0920bdb8a9a18f6ea9c52cf5028e9d"} Nov 29 04:55:50 crc kubenswrapper[4799]: I1129 04:55:50.535592 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-76859cbf9f-kdxtl" event={"ID":"f14d7038-795e-4f8a-8ce7-81e64c86137b","Type":"ContainerStarted","Data":"8ac82f8cf7b94bae207a574b62503707d3aea8a4a1011e7751c6581becc1f106"} Nov 29 04:55:50 crc kubenswrapper[4799]: I1129 04:55:50.537641 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-btd8j" event={"ID":"515c81c7-10e3-4724-aafd-42431946c2c2","Type":"ContainerStarted","Data":"a1e80e70152500fed314f16d33879352b8ab7671cbd3309602c3c9653156594b"} Nov 29 04:55:50 crc kubenswrapper[4799]: I1129 04:55:50.543512 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6495665bf6-xn8f7" event={"ID":"bdd214a8-9094-42ef-b2c6-2452b59596f7","Type":"ContainerStarted","Data":"4281ed4377c37de4d122f1203ab93a4eee72a38853031074e51f4a9f5647406b"} Nov 29 04:55:50 crc kubenswrapper[4799]: I1129 04:55:50.554069 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"82c42edf-7b25-411f-b00f-95ed3e74bede","Type":"ContainerStarted","Data":"9c18a41315f857ee4061e7705406ba0dbb294b779bc6c834ee67c8a30824060b"} Nov 29 04:55:50 crc kubenswrapper[4799]: I1129 04:55:50.554284 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="82c42edf-7b25-411f-b00f-95ed3e74bede" containerName="ceilometer-central-agent" containerID="cri-o://19f81f19bd1e89c5024a9fd7c16e1c64a6b91117546b1cfd8f819f63de7fdfb1" gracePeriod=30 Nov 29 04:55:50 crc kubenswrapper[4799]: I1129 04:55:50.554545 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 29 04:55:50 crc kubenswrapper[4799]: I1129 04:55:50.554592 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="82c42edf-7b25-411f-b00f-95ed3e74bede" containerName="proxy-httpd" containerID="cri-o://9c18a41315f857ee4061e7705406ba0dbb294b779bc6c834ee67c8a30824060b" gracePeriod=30 Nov 29 04:55:50 crc kubenswrapper[4799]: I1129 04:55:50.554640 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="82c42edf-7b25-411f-b00f-95ed3e74bede" containerName="sg-core" containerID="cri-o://e5541ed59a78cb4ea523c9cbac248a33c6b6584def87b5d8ecd5bac9f2095cd0" gracePeriod=30 Nov 29 04:55:50 crc kubenswrapper[4799]: I1129 04:55:50.555102 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="82c42edf-7b25-411f-b00f-95ed3e74bede" containerName="ceilometer-notification-agent" containerID="cri-o://029c5b8b863f82633a82b4acc01bab9cc121c4cc9a483238bf2cf3a6bd02cb66" gracePeriod=30 Nov 29 04:55:50 crc kubenswrapper[4799]: I1129 04:55:50.565221 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-669f887b5-74klb" event={"ID":"122be3fa-bc40-45f0-808d-fece3e876c43","Type":"ContainerStarted","Data":"f53d60b388271ab104549ac94148f6f0fa669466616698b29f9bf962e83234e1"} Nov 29 04:55:50 crc kubenswrapper[4799]: I1129 04:55:50.578450 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-btd8j" podStartSLOduration=3.862540564 podStartE2EDuration="54.578427344s" podCreationTimestamp="2025-11-29 04:54:56 +0000 UTC" firstStartedPulling="2025-11-29 04:54:58.112863489 +0000 UTC m=+973.755793889" lastFinishedPulling="2025-11-29 04:55:48.828750269 +0000 UTC m=+1024.471680669" observedRunningTime="2025-11-29 04:55:50.571701439 +0000 UTC m=+1026.214631839" watchObservedRunningTime="2025-11-29 04:55:50.578427344 +0000 UTC m=+1026.221357744" Nov 29 04:55:50 crc kubenswrapper[4799]: I1129 04:55:50.600299 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b946d459c-5twn6"] Nov 29 04:55:50 crc kubenswrapper[4799]: I1129 04:55:50.610206 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7b946d459c-5twn6"] Nov 29 04:55:50 crc kubenswrapper[4799]: I1129 04:55:50.618415 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.955219965 podStartE2EDuration="53.618394208s" podCreationTimestamp="2025-11-29 04:54:57 +0000 UTC" firstStartedPulling="2025-11-29 04:54:58.113905715 +0000 UTC m=+973.756836115" lastFinishedPulling="2025-11-29 04:55:48.777079958 +0000 UTC m=+1024.420010358" observedRunningTime="2025-11-29 04:55:50.616593145 +0000 UTC m=+1026.259523555" watchObservedRunningTime="2025-11-29 04:55:50.618394208 +0000 UTC m=+1026.261324608" Nov 29 04:55:50 crc kubenswrapper[4799]: I1129 04:55:50.674547 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4" path="/var/lib/kubelet/pods/5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4/volumes" Nov 29 04:55:51 crc kubenswrapper[4799]: I1129 04:55:51.117129 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-fdfbddb56-lhh8r"] Nov 29 04:55:51 crc kubenswrapper[4799]: E1129 04:55:51.117652 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4" containerName="init" Nov 29 04:55:51 crc kubenswrapper[4799]: I1129 04:55:51.117669 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4" containerName="init" Nov 29 04:55:51 crc kubenswrapper[4799]: E1129 04:55:51.117745 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4" containerName="dnsmasq-dns" Nov 29 04:55:51 crc kubenswrapper[4799]: I1129 04:55:51.117757 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4" containerName="dnsmasq-dns" Nov 29 04:55:51 crc kubenswrapper[4799]: I1129 04:55:51.118561 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f0b2f98-90ce-4ca1-84b5-6d88e609a9e4" containerName="dnsmasq-dns" Nov 29 04:55:51 crc kubenswrapper[4799]: I1129 04:55:51.120871 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-fdfbddb56-lhh8r" Nov 29 04:55:51 crc kubenswrapper[4799]: I1129 04:55:51.125673 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Nov 29 04:55:51 crc kubenswrapper[4799]: I1129 04:55:51.133514 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Nov 29 04:55:51 crc kubenswrapper[4799]: I1129 04:55:51.147121 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-fdfbddb56-lhh8r"] Nov 29 04:55:51 crc kubenswrapper[4799]: I1129 04:55:51.214878 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wz8jk\" (UniqueName: \"kubernetes.io/projected/ddd039b3-89fb-4515-a633-e4c8c23bc25c-kube-api-access-wz8jk\") pod \"barbican-api-fdfbddb56-lhh8r\" (UID: \"ddd039b3-89fb-4515-a633-e4c8c23bc25c\") " pod="openstack/barbican-api-fdfbddb56-lhh8r" Nov 29 04:55:51 crc kubenswrapper[4799]: I1129 04:55:51.214944 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddd039b3-89fb-4515-a633-e4c8c23bc25c-public-tls-certs\") pod \"barbican-api-fdfbddb56-lhh8r\" (UID: \"ddd039b3-89fb-4515-a633-e4c8c23bc25c\") " pod="openstack/barbican-api-fdfbddb56-lhh8r" Nov 29 04:55:51 crc kubenswrapper[4799]: I1129 04:55:51.214999 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ddd039b3-89fb-4515-a633-e4c8c23bc25c-config-data-custom\") pod \"barbican-api-fdfbddb56-lhh8r\" (UID: \"ddd039b3-89fb-4515-a633-e4c8c23bc25c\") " pod="openstack/barbican-api-fdfbddb56-lhh8r" Nov 29 04:55:51 crc kubenswrapper[4799]: I1129 04:55:51.215354 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddd039b3-89fb-4515-a633-e4c8c23bc25c-config-data\") pod \"barbican-api-fdfbddb56-lhh8r\" (UID: \"ddd039b3-89fb-4515-a633-e4c8c23bc25c\") " pod="openstack/barbican-api-fdfbddb56-lhh8r" Nov 29 04:55:51 crc kubenswrapper[4799]: I1129 04:55:51.215540 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddd039b3-89fb-4515-a633-e4c8c23bc25c-combined-ca-bundle\") pod \"barbican-api-fdfbddb56-lhh8r\" (UID: \"ddd039b3-89fb-4515-a633-e4c8c23bc25c\") " pod="openstack/barbican-api-fdfbddb56-lhh8r" Nov 29 04:55:51 crc kubenswrapper[4799]: I1129 04:55:51.215739 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddd039b3-89fb-4515-a633-e4c8c23bc25c-internal-tls-certs\") pod \"barbican-api-fdfbddb56-lhh8r\" (UID: \"ddd039b3-89fb-4515-a633-e4c8c23bc25c\") " pod="openstack/barbican-api-fdfbddb56-lhh8r" Nov 29 04:55:51 crc kubenswrapper[4799]: I1129 04:55:51.215846 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ddd039b3-89fb-4515-a633-e4c8c23bc25c-logs\") pod \"barbican-api-fdfbddb56-lhh8r\" (UID: \"ddd039b3-89fb-4515-a633-e4c8c23bc25c\") " pod="openstack/barbican-api-fdfbddb56-lhh8r" Nov 29 04:55:51 crc kubenswrapper[4799]: I1129 04:55:51.318079 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddd039b3-89fb-4515-a633-e4c8c23bc25c-config-data\") pod \"barbican-api-fdfbddb56-lhh8r\" (UID: \"ddd039b3-89fb-4515-a633-e4c8c23bc25c\") " pod="openstack/barbican-api-fdfbddb56-lhh8r" Nov 29 04:55:51 crc kubenswrapper[4799]: I1129 04:55:51.318165 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddd039b3-89fb-4515-a633-e4c8c23bc25c-combined-ca-bundle\") pod \"barbican-api-fdfbddb56-lhh8r\" (UID: \"ddd039b3-89fb-4515-a633-e4c8c23bc25c\") " pod="openstack/barbican-api-fdfbddb56-lhh8r" Nov 29 04:55:51 crc kubenswrapper[4799]: I1129 04:55:51.318224 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddd039b3-89fb-4515-a633-e4c8c23bc25c-internal-tls-certs\") pod \"barbican-api-fdfbddb56-lhh8r\" (UID: \"ddd039b3-89fb-4515-a633-e4c8c23bc25c\") " pod="openstack/barbican-api-fdfbddb56-lhh8r" Nov 29 04:55:51 crc kubenswrapper[4799]: I1129 04:55:51.318250 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ddd039b3-89fb-4515-a633-e4c8c23bc25c-logs\") pod \"barbican-api-fdfbddb56-lhh8r\" (UID: \"ddd039b3-89fb-4515-a633-e4c8c23bc25c\") " pod="openstack/barbican-api-fdfbddb56-lhh8r" Nov 29 04:55:51 crc kubenswrapper[4799]: I1129 04:55:51.318302 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wz8jk\" (UniqueName: \"kubernetes.io/projected/ddd039b3-89fb-4515-a633-e4c8c23bc25c-kube-api-access-wz8jk\") pod \"barbican-api-fdfbddb56-lhh8r\" (UID: \"ddd039b3-89fb-4515-a633-e4c8c23bc25c\") " pod="openstack/barbican-api-fdfbddb56-lhh8r" Nov 29 04:55:51 crc kubenswrapper[4799]: I1129 04:55:51.318324 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddd039b3-89fb-4515-a633-e4c8c23bc25c-public-tls-certs\") pod \"barbican-api-fdfbddb56-lhh8r\" (UID: \"ddd039b3-89fb-4515-a633-e4c8c23bc25c\") " pod="openstack/barbican-api-fdfbddb56-lhh8r" Nov 29 04:55:51 crc kubenswrapper[4799]: I1129 04:55:51.318370 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ddd039b3-89fb-4515-a633-e4c8c23bc25c-config-data-custom\") pod \"barbican-api-fdfbddb56-lhh8r\" (UID: \"ddd039b3-89fb-4515-a633-e4c8c23bc25c\") " pod="openstack/barbican-api-fdfbddb56-lhh8r" Nov 29 04:55:51 crc kubenswrapper[4799]: I1129 04:55:51.319227 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ddd039b3-89fb-4515-a633-e4c8c23bc25c-logs\") pod \"barbican-api-fdfbddb56-lhh8r\" (UID: \"ddd039b3-89fb-4515-a633-e4c8c23bc25c\") " pod="openstack/barbican-api-fdfbddb56-lhh8r" Nov 29 04:55:51 crc kubenswrapper[4799]: I1129 04:55:51.324772 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ddd039b3-89fb-4515-a633-e4c8c23bc25c-config-data-custom\") pod \"barbican-api-fdfbddb56-lhh8r\" (UID: \"ddd039b3-89fb-4515-a633-e4c8c23bc25c\") " pod="openstack/barbican-api-fdfbddb56-lhh8r" Nov 29 04:55:51 crc kubenswrapper[4799]: I1129 04:55:51.325540 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddd039b3-89fb-4515-a633-e4c8c23bc25c-combined-ca-bundle\") pod \"barbican-api-fdfbddb56-lhh8r\" (UID: \"ddd039b3-89fb-4515-a633-e4c8c23bc25c\") " pod="openstack/barbican-api-fdfbddb56-lhh8r" Nov 29 04:55:51 crc kubenswrapper[4799]: I1129 04:55:51.326036 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddd039b3-89fb-4515-a633-e4c8c23bc25c-config-data\") pod \"barbican-api-fdfbddb56-lhh8r\" (UID: \"ddd039b3-89fb-4515-a633-e4c8c23bc25c\") " pod="openstack/barbican-api-fdfbddb56-lhh8r" Nov 29 04:55:51 crc kubenswrapper[4799]: I1129 04:55:51.327560 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddd039b3-89fb-4515-a633-e4c8c23bc25c-public-tls-certs\") pod \"barbican-api-fdfbddb56-lhh8r\" (UID: \"ddd039b3-89fb-4515-a633-e4c8c23bc25c\") " pod="openstack/barbican-api-fdfbddb56-lhh8r" Nov 29 04:55:51 crc kubenswrapper[4799]: I1129 04:55:51.333703 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddd039b3-89fb-4515-a633-e4c8c23bc25c-internal-tls-certs\") pod \"barbican-api-fdfbddb56-lhh8r\" (UID: \"ddd039b3-89fb-4515-a633-e4c8c23bc25c\") " pod="openstack/barbican-api-fdfbddb56-lhh8r" Nov 29 04:55:51 crc kubenswrapper[4799]: I1129 04:55:51.346021 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wz8jk\" (UniqueName: \"kubernetes.io/projected/ddd039b3-89fb-4515-a633-e4c8c23bc25c-kube-api-access-wz8jk\") pod \"barbican-api-fdfbddb56-lhh8r\" (UID: \"ddd039b3-89fb-4515-a633-e4c8c23bc25c\") " pod="openstack/barbican-api-fdfbddb56-lhh8r" Nov 29 04:55:51 crc kubenswrapper[4799]: I1129 04:55:51.481234 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-fdfbddb56-lhh8r" Nov 29 04:55:51 crc kubenswrapper[4799]: I1129 04:55:51.580744 4799 generic.go:334] "Generic (PLEG): container finished" podID="82c42edf-7b25-411f-b00f-95ed3e74bede" containerID="9c18a41315f857ee4061e7705406ba0dbb294b779bc6c834ee67c8a30824060b" exitCode=0 Nov 29 04:55:51 crc kubenswrapper[4799]: I1129 04:55:51.580811 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"82c42edf-7b25-411f-b00f-95ed3e74bede","Type":"ContainerDied","Data":"9c18a41315f857ee4061e7705406ba0dbb294b779bc6c834ee67c8a30824060b"} Nov 29 04:55:51 crc kubenswrapper[4799]: I1129 04:55:51.580875 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"82c42edf-7b25-411f-b00f-95ed3e74bede","Type":"ContainerDied","Data":"e5541ed59a78cb4ea523c9cbac248a33c6b6584def87b5d8ecd5bac9f2095cd0"} Nov 29 04:55:51 crc kubenswrapper[4799]: I1129 04:55:51.580784 4799 generic.go:334] "Generic (PLEG): container finished" podID="82c42edf-7b25-411f-b00f-95ed3e74bede" containerID="e5541ed59a78cb4ea523c9cbac248a33c6b6584def87b5d8ecd5bac9f2095cd0" exitCode=2 Nov 29 04:55:51 crc kubenswrapper[4799]: I1129 04:55:51.580899 4799 generic.go:334] "Generic (PLEG): container finished" podID="82c42edf-7b25-411f-b00f-95ed3e74bede" containerID="19f81f19bd1e89c5024a9fd7c16e1c64a6b91117546b1cfd8f819f63de7fdfb1" exitCode=0 Nov 29 04:55:51 crc kubenswrapper[4799]: I1129 04:55:51.580964 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"82c42edf-7b25-411f-b00f-95ed3e74bede","Type":"ContainerDied","Data":"19f81f19bd1e89c5024a9fd7c16e1c64a6b91117546b1cfd8f819f63de7fdfb1"} Nov 29 04:55:51 crc kubenswrapper[4799]: I1129 04:55:51.583707 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-669f887b5-74klb" event={"ID":"122be3fa-bc40-45f0-808d-fece3e876c43","Type":"ContainerStarted","Data":"b4d7a5e1602050e920334936fb7c3eee005bda991bd5f8f6fc7fb0ba1e09c31d"} Nov 29 04:55:51 crc kubenswrapper[4799]: I1129 04:55:51.585502 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-669f887b5-74klb" Nov 29 04:55:51 crc kubenswrapper[4799]: I1129 04:55:51.589551 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb684768f-x97jl" event={"ID":"0ae506e6-d401-4fa6-b4ce-8db525ef694e","Type":"ContainerStarted","Data":"98920f3f0ac9e5a6e35fcc7c3647d9f155b455fe2b3f70acf845c665ed352fb3"} Nov 29 04:55:51 crc kubenswrapper[4799]: I1129 04:55:51.589830 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6bb684768f-x97jl" Nov 29 04:55:51 crc kubenswrapper[4799]: I1129 04:55:51.592395 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6944bdb4b-6bt2h" event={"ID":"e964f0f8-6843-4a3c-917b-69372833af7a","Type":"ContainerStarted","Data":"8c6707e699584a656f44a5d76623c1304572bf57d4ffd926e1a45018d4470029"} Nov 29 04:55:51 crc kubenswrapper[4799]: I1129 04:55:51.592653 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6944bdb4b-6bt2h" Nov 29 04:55:51 crc kubenswrapper[4799]: I1129 04:55:51.616088 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-669f887b5-74klb" podStartSLOduration=9.616062394 podStartE2EDuration="9.616062394s" podCreationTimestamp="2025-11-29 04:55:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:55:51.612264961 +0000 UTC m=+1027.255195361" watchObservedRunningTime="2025-11-29 04:55:51.616062394 +0000 UTC m=+1027.258992794" Nov 29 04:55:51 crc kubenswrapper[4799]: I1129 04:55:51.642666 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6bb684768f-x97jl" podStartSLOduration=3.642649122 podStartE2EDuration="3.642649122s" podCreationTimestamp="2025-11-29 04:55:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:55:51.641564896 +0000 UTC m=+1027.284495296" watchObservedRunningTime="2025-11-29 04:55:51.642649122 +0000 UTC m=+1027.285579522" Nov 29 04:55:52 crc kubenswrapper[4799]: I1129 04:55:52.518000 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6944bdb4b-6bt2h" podStartSLOduration=4.517971931 podStartE2EDuration="4.517971931s" podCreationTimestamp="2025-11-29 04:55:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:55:51.6671627 +0000 UTC m=+1027.310093100" watchObservedRunningTime="2025-11-29 04:55:52.517971931 +0000 UTC m=+1028.160902341" Nov 29 04:55:52 crc kubenswrapper[4799]: I1129 04:55:52.519176 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-fdfbddb56-lhh8r"] Nov 29 04:55:52 crc kubenswrapper[4799]: I1129 04:55:52.610577 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-76859cbf9f-kdxtl" event={"ID":"f14d7038-795e-4f8a-8ce7-81e64c86137b","Type":"ContainerStarted","Data":"50b08d05011c79d0980a45a9cb6602470b4ffded720ec343e5c5ceb3382b6c1c"} Nov 29 04:55:52 crc kubenswrapper[4799]: I1129 04:55:52.613768 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6495665bf6-xn8f7" event={"ID":"bdd214a8-9094-42ef-b2c6-2452b59596f7","Type":"ContainerStarted","Data":"65344b2857fd3dd78d35896664b13b99d98a005594610ebfa5116af8a13e0b11"} Nov 29 04:55:52 crc kubenswrapper[4799]: I1129 04:55:52.617697 4799 generic.go:334] "Generic (PLEG): container finished" podID="82c42edf-7b25-411f-b00f-95ed3e74bede" containerID="029c5b8b863f82633a82b4acc01bab9cc121c4cc9a483238bf2cf3a6bd02cb66" exitCode=0 Nov 29 04:55:52 crc kubenswrapper[4799]: I1129 04:55:52.617770 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"82c42edf-7b25-411f-b00f-95ed3e74bede","Type":"ContainerDied","Data":"029c5b8b863f82633a82b4acc01bab9cc121c4cc9a483238bf2cf3a6bd02cb66"} Nov 29 04:55:52 crc kubenswrapper[4799]: I1129 04:55:52.620052 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-fdfbddb56-lhh8r" event={"ID":"ddd039b3-89fb-4515-a633-e4c8c23bc25c","Type":"ContainerStarted","Data":"0a2669fb0b32fad6500bb63cbc8e347857b5a72b07e1bc4d2d8c243b217496bf"} Nov 29 04:55:52 crc kubenswrapper[4799]: I1129 04:55:52.621340 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6944bdb4b-6bt2h" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.034580 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.163849 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82c42edf-7b25-411f-b00f-95ed3e74bede-scripts\") pod \"82c42edf-7b25-411f-b00f-95ed3e74bede\" (UID: \"82c42edf-7b25-411f-b00f-95ed3e74bede\") " Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.164028 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zld5m\" (UniqueName: \"kubernetes.io/projected/82c42edf-7b25-411f-b00f-95ed3e74bede-kube-api-access-zld5m\") pod \"82c42edf-7b25-411f-b00f-95ed3e74bede\" (UID: \"82c42edf-7b25-411f-b00f-95ed3e74bede\") " Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.164108 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/82c42edf-7b25-411f-b00f-95ed3e74bede-log-httpd\") pod \"82c42edf-7b25-411f-b00f-95ed3e74bede\" (UID: \"82c42edf-7b25-411f-b00f-95ed3e74bede\") " Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.164231 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82c42edf-7b25-411f-b00f-95ed3e74bede-combined-ca-bundle\") pod \"82c42edf-7b25-411f-b00f-95ed3e74bede\" (UID: \"82c42edf-7b25-411f-b00f-95ed3e74bede\") " Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.164266 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/82c42edf-7b25-411f-b00f-95ed3e74bede-sg-core-conf-yaml\") pod \"82c42edf-7b25-411f-b00f-95ed3e74bede\" (UID: \"82c42edf-7b25-411f-b00f-95ed3e74bede\") " Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.164294 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82c42edf-7b25-411f-b00f-95ed3e74bede-config-data\") pod \"82c42edf-7b25-411f-b00f-95ed3e74bede\" (UID: \"82c42edf-7b25-411f-b00f-95ed3e74bede\") " Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.164324 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/82c42edf-7b25-411f-b00f-95ed3e74bede-run-httpd\") pod \"82c42edf-7b25-411f-b00f-95ed3e74bede\" (UID: \"82c42edf-7b25-411f-b00f-95ed3e74bede\") " Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.165053 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82c42edf-7b25-411f-b00f-95ed3e74bede-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "82c42edf-7b25-411f-b00f-95ed3e74bede" (UID: "82c42edf-7b25-411f-b00f-95ed3e74bede"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.169502 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82c42edf-7b25-411f-b00f-95ed3e74bede-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "82c42edf-7b25-411f-b00f-95ed3e74bede" (UID: "82c42edf-7b25-411f-b00f-95ed3e74bede"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.176125 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82c42edf-7b25-411f-b00f-95ed3e74bede-kube-api-access-zld5m" (OuterVolumeSpecName: "kube-api-access-zld5m") pod "82c42edf-7b25-411f-b00f-95ed3e74bede" (UID: "82c42edf-7b25-411f-b00f-95ed3e74bede"). InnerVolumeSpecName "kube-api-access-zld5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.176262 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82c42edf-7b25-411f-b00f-95ed3e74bede-scripts" (OuterVolumeSpecName: "scripts") pod "82c42edf-7b25-411f-b00f-95ed3e74bede" (UID: "82c42edf-7b25-411f-b00f-95ed3e74bede"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.210510 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82c42edf-7b25-411f-b00f-95ed3e74bede-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "82c42edf-7b25-411f-b00f-95ed3e74bede" (UID: "82c42edf-7b25-411f-b00f-95ed3e74bede"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.266424 4799 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/82c42edf-7b25-411f-b00f-95ed3e74bede-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.266721 4799 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/82c42edf-7b25-411f-b00f-95ed3e74bede-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.266782 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82c42edf-7b25-411f-b00f-95ed3e74bede-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.266860 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zld5m\" (UniqueName: \"kubernetes.io/projected/82c42edf-7b25-411f-b00f-95ed3e74bede-kube-api-access-zld5m\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.266908 4799 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/82c42edf-7b25-411f-b00f-95ed3e74bede-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.270953 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82c42edf-7b25-411f-b00f-95ed3e74bede-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "82c42edf-7b25-411f-b00f-95ed3e74bede" (UID: "82c42edf-7b25-411f-b00f-95ed3e74bede"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.313239 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82c42edf-7b25-411f-b00f-95ed3e74bede-config-data" (OuterVolumeSpecName: "config-data") pod "82c42edf-7b25-411f-b00f-95ed3e74bede" (UID: "82c42edf-7b25-411f-b00f-95ed3e74bede"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.368880 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82c42edf-7b25-411f-b00f-95ed3e74bede-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.369157 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82c42edf-7b25-411f-b00f-95ed3e74bede-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.635076 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-76859cbf9f-kdxtl" event={"ID":"f14d7038-795e-4f8a-8ce7-81e64c86137b","Type":"ContainerStarted","Data":"a79977a114808b4cdd093c03f2b4da1fb074a3af6910e2c2832570c07691ed45"} Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.641820 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6495665bf6-xn8f7" event={"ID":"bdd214a8-9094-42ef-b2c6-2452b59596f7","Type":"ContainerStarted","Data":"5305aabedd06d0884a1d3c8310f69e1077bdd7246884b350b554398f6b550401"} Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.647434 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"82c42edf-7b25-411f-b00f-95ed3e74bede","Type":"ContainerDied","Data":"107ffbdcfced5b102126423c5623d97eae4be041870337b29984647a277720a5"} Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.647716 4799 scope.go:117] "RemoveContainer" containerID="9c18a41315f857ee4061e7705406ba0dbb294b779bc6c834ee67c8a30824060b" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.648079 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.664949 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-fdfbddb56-lhh8r" event={"ID":"ddd039b3-89fb-4515-a633-e4c8c23bc25c","Type":"ContainerStarted","Data":"b39d5e7cf8c691db774fe9b1667d178375fcd70766bd6419cd902d449dc1c852"} Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.664996 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-fdfbddb56-lhh8r" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.665005 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-fdfbddb56-lhh8r" event={"ID":"ddd039b3-89fb-4515-a633-e4c8c23bc25c","Type":"ContainerStarted","Data":"8613f139a1f090e8e3e42cbff0aa3b281a1f5e10260a05f3df47abaeefd4a197"} Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.665427 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-fdfbddb56-lhh8r" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.704646 4799 scope.go:117] "RemoveContainer" containerID="e5541ed59a78cb4ea523c9cbac248a33c6b6584def87b5d8ecd5bac9f2095cd0" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.706665 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-6495665bf6-xn8f7" podStartSLOduration=4.284101983 podStartE2EDuration="6.706636427s" podCreationTimestamp="2025-11-29 04:55:47 +0000 UTC" firstStartedPulling="2025-11-29 04:55:49.655085772 +0000 UTC m=+1025.298016172" lastFinishedPulling="2025-11-29 04:55:52.077620216 +0000 UTC m=+1027.720550616" observedRunningTime="2025-11-29 04:55:53.695426043 +0000 UTC m=+1029.338356433" watchObservedRunningTime="2025-11-29 04:55:53.706636427 +0000 UTC m=+1029.349566837" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.710747 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-76859cbf9f-kdxtl" podStartSLOduration=4.316609867 podStartE2EDuration="6.710734207s" podCreationTimestamp="2025-11-29 04:55:47 +0000 UTC" firstStartedPulling="2025-11-29 04:55:49.680236376 +0000 UTC m=+1025.323166776" lastFinishedPulling="2025-11-29 04:55:52.074360716 +0000 UTC m=+1027.717291116" observedRunningTime="2025-11-29 04:55:53.662194782 +0000 UTC m=+1029.305125182" watchObservedRunningTime="2025-11-29 04:55:53.710734207 +0000 UTC m=+1029.353664607" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.726994 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.735488 4799 scope.go:117] "RemoveContainer" containerID="029c5b8b863f82633a82b4acc01bab9cc121c4cc9a483238bf2cf3a6bd02cb66" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.739707 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.760028 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-fdfbddb56-lhh8r" podStartSLOduration=3.760002269 podStartE2EDuration="3.760002269s" podCreationTimestamp="2025-11-29 04:55:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:55:53.755726174 +0000 UTC m=+1029.398656574" watchObservedRunningTime="2025-11-29 04:55:53.760002269 +0000 UTC m=+1029.402932669" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.774093 4799 scope.go:117] "RemoveContainer" containerID="19f81f19bd1e89c5024a9fd7c16e1c64a6b91117546b1cfd8f819f63de7fdfb1" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.777596 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 29 04:55:53 crc kubenswrapper[4799]: E1129 04:55:53.778167 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82c42edf-7b25-411f-b00f-95ed3e74bede" containerName="ceilometer-central-agent" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.778192 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="82c42edf-7b25-411f-b00f-95ed3e74bede" containerName="ceilometer-central-agent" Nov 29 04:55:53 crc kubenswrapper[4799]: E1129 04:55:53.778222 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82c42edf-7b25-411f-b00f-95ed3e74bede" containerName="ceilometer-notification-agent" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.778231 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="82c42edf-7b25-411f-b00f-95ed3e74bede" containerName="ceilometer-notification-agent" Nov 29 04:55:53 crc kubenswrapper[4799]: E1129 04:55:53.778241 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82c42edf-7b25-411f-b00f-95ed3e74bede" containerName="sg-core" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.778248 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="82c42edf-7b25-411f-b00f-95ed3e74bede" containerName="sg-core" Nov 29 04:55:53 crc kubenswrapper[4799]: E1129 04:55:53.778270 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82c42edf-7b25-411f-b00f-95ed3e74bede" containerName="proxy-httpd" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.778301 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="82c42edf-7b25-411f-b00f-95ed3e74bede" containerName="proxy-httpd" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.778489 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="82c42edf-7b25-411f-b00f-95ed3e74bede" containerName="ceilometer-central-agent" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.778522 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="82c42edf-7b25-411f-b00f-95ed3e74bede" containerName="proxy-httpd" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.778535 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="82c42edf-7b25-411f-b00f-95ed3e74bede" containerName="sg-core" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.778551 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="82c42edf-7b25-411f-b00f-95ed3e74bede" containerName="ceilometer-notification-agent" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.780351 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.783895 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.825133 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.825470 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.877963 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/de975367-f439-4788-8ddb-b9f8af482589-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"de975367-f439-4788-8ddb-b9f8af482589\") " pod="openstack/ceilometer-0" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.879152 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de975367-f439-4788-8ddb-b9f8af482589-scripts\") pod \"ceilometer-0\" (UID: \"de975367-f439-4788-8ddb-b9f8af482589\") " pod="openstack/ceilometer-0" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.879291 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2d22\" (UniqueName: \"kubernetes.io/projected/de975367-f439-4788-8ddb-b9f8af482589-kube-api-access-f2d22\") pod \"ceilometer-0\" (UID: \"de975367-f439-4788-8ddb-b9f8af482589\") " pod="openstack/ceilometer-0" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.879450 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de975367-f439-4788-8ddb-b9f8af482589-config-data\") pod \"ceilometer-0\" (UID: \"de975367-f439-4788-8ddb-b9f8af482589\") " pod="openstack/ceilometer-0" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.879514 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de975367-f439-4788-8ddb-b9f8af482589-run-httpd\") pod \"ceilometer-0\" (UID: \"de975367-f439-4788-8ddb-b9f8af482589\") " pod="openstack/ceilometer-0" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.879603 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de975367-f439-4788-8ddb-b9f8af482589-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"de975367-f439-4788-8ddb-b9f8af482589\") " pod="openstack/ceilometer-0" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.879673 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de975367-f439-4788-8ddb-b9f8af482589-log-httpd\") pod \"ceilometer-0\" (UID: \"de975367-f439-4788-8ddb-b9f8af482589\") " pod="openstack/ceilometer-0" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.981528 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/de975367-f439-4788-8ddb-b9f8af482589-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"de975367-f439-4788-8ddb-b9f8af482589\") " pod="openstack/ceilometer-0" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.981611 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de975367-f439-4788-8ddb-b9f8af482589-scripts\") pod \"ceilometer-0\" (UID: \"de975367-f439-4788-8ddb-b9f8af482589\") " pod="openstack/ceilometer-0" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.981631 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2d22\" (UniqueName: \"kubernetes.io/projected/de975367-f439-4788-8ddb-b9f8af482589-kube-api-access-f2d22\") pod \"ceilometer-0\" (UID: \"de975367-f439-4788-8ddb-b9f8af482589\") " pod="openstack/ceilometer-0" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.981697 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de975367-f439-4788-8ddb-b9f8af482589-config-data\") pod \"ceilometer-0\" (UID: \"de975367-f439-4788-8ddb-b9f8af482589\") " pod="openstack/ceilometer-0" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.981720 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de975367-f439-4788-8ddb-b9f8af482589-run-httpd\") pod \"ceilometer-0\" (UID: \"de975367-f439-4788-8ddb-b9f8af482589\") " pod="openstack/ceilometer-0" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.981750 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de975367-f439-4788-8ddb-b9f8af482589-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"de975367-f439-4788-8ddb-b9f8af482589\") " pod="openstack/ceilometer-0" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.981770 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de975367-f439-4788-8ddb-b9f8af482589-log-httpd\") pod \"ceilometer-0\" (UID: \"de975367-f439-4788-8ddb-b9f8af482589\") " pod="openstack/ceilometer-0" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.982360 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de975367-f439-4788-8ddb-b9f8af482589-log-httpd\") pod \"ceilometer-0\" (UID: \"de975367-f439-4788-8ddb-b9f8af482589\") " pod="openstack/ceilometer-0" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.983389 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de975367-f439-4788-8ddb-b9f8af482589-run-httpd\") pod \"ceilometer-0\" (UID: \"de975367-f439-4788-8ddb-b9f8af482589\") " pod="openstack/ceilometer-0" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.990241 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/de975367-f439-4788-8ddb-b9f8af482589-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"de975367-f439-4788-8ddb-b9f8af482589\") " pod="openstack/ceilometer-0" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.993625 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de975367-f439-4788-8ddb-b9f8af482589-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"de975367-f439-4788-8ddb-b9f8af482589\") " pod="openstack/ceilometer-0" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.994014 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de975367-f439-4788-8ddb-b9f8af482589-config-data\") pod \"ceilometer-0\" (UID: \"de975367-f439-4788-8ddb-b9f8af482589\") " pod="openstack/ceilometer-0" Nov 29 04:55:53 crc kubenswrapper[4799]: I1129 04:55:53.997528 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de975367-f439-4788-8ddb-b9f8af482589-scripts\") pod \"ceilometer-0\" (UID: \"de975367-f439-4788-8ddb-b9f8af482589\") " pod="openstack/ceilometer-0" Nov 29 04:55:54 crc kubenswrapper[4799]: I1129 04:55:54.004421 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2d22\" (UniqueName: \"kubernetes.io/projected/de975367-f439-4788-8ddb-b9f8af482589-kube-api-access-f2d22\") pod \"ceilometer-0\" (UID: \"de975367-f439-4788-8ddb-b9f8af482589\") " pod="openstack/ceilometer-0" Nov 29 04:55:54 crc kubenswrapper[4799]: I1129 04:55:54.132447 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 04:55:54 crc kubenswrapper[4799]: I1129 04:55:54.675135 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82c42edf-7b25-411f-b00f-95ed3e74bede" path="/var/lib/kubelet/pods/82c42edf-7b25-411f-b00f-95ed3e74bede/volumes" Nov 29 04:55:55 crc kubenswrapper[4799]: I1129 04:55:55.149823 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 29 04:55:55 crc kubenswrapper[4799]: I1129 04:55:55.716741 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de975367-f439-4788-8ddb-b9f8af482589","Type":"ContainerStarted","Data":"3aa3cf1b4350391a08cc4f094520d74e14370c0def867d3fd71492f1dc600631"} Nov 29 04:55:56 crc kubenswrapper[4799]: I1129 04:55:56.730772 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de975367-f439-4788-8ddb-b9f8af482589","Type":"ContainerStarted","Data":"59ea15e9044a8c7f1247d7fb32c1ad6ab35c795668363a83216a4a7e201564c0"} Nov 29 04:55:56 crc kubenswrapper[4799]: I1129 04:55:56.734183 4799 generic.go:334] "Generic (PLEG): container finished" podID="515c81c7-10e3-4724-aafd-42431946c2c2" containerID="a1e80e70152500fed314f16d33879352b8ab7671cbd3309602c3c9653156594b" exitCode=0 Nov 29 04:55:56 crc kubenswrapper[4799]: I1129 04:55:56.734233 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-btd8j" event={"ID":"515c81c7-10e3-4724-aafd-42431946c2c2","Type":"ContainerDied","Data":"a1e80e70152500fed314f16d33879352b8ab7671cbd3309602c3c9653156594b"} Nov 29 04:55:57 crc kubenswrapper[4799]: I1129 04:55:57.805960 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de975367-f439-4788-8ddb-b9f8af482589","Type":"ContainerStarted","Data":"fbd3059919aa4e1d1d2943fd052b2861ed1b731729d22d5383486cc878702a33"} Nov 29 04:55:58 crc kubenswrapper[4799]: I1129 04:55:58.192619 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-btd8j" Nov 29 04:55:58 crc kubenswrapper[4799]: I1129 04:55:58.289292 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/515c81c7-10e3-4724-aafd-42431946c2c2-etc-machine-id\") pod \"515c81c7-10e3-4724-aafd-42431946c2c2\" (UID: \"515c81c7-10e3-4724-aafd-42431946c2c2\") " Nov 29 04:55:58 crc kubenswrapper[4799]: I1129 04:55:58.289438 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/515c81c7-10e3-4724-aafd-42431946c2c2-db-sync-config-data\") pod \"515c81c7-10e3-4724-aafd-42431946c2c2\" (UID: \"515c81c7-10e3-4724-aafd-42431946c2c2\") " Nov 29 04:55:58 crc kubenswrapper[4799]: I1129 04:55:58.289492 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/515c81c7-10e3-4724-aafd-42431946c2c2-scripts\") pod \"515c81c7-10e3-4724-aafd-42431946c2c2\" (UID: \"515c81c7-10e3-4724-aafd-42431946c2c2\") " Nov 29 04:55:58 crc kubenswrapper[4799]: I1129 04:55:58.289544 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-shprr\" (UniqueName: \"kubernetes.io/projected/515c81c7-10e3-4724-aafd-42431946c2c2-kube-api-access-shprr\") pod \"515c81c7-10e3-4724-aafd-42431946c2c2\" (UID: \"515c81c7-10e3-4724-aafd-42431946c2c2\") " Nov 29 04:55:58 crc kubenswrapper[4799]: I1129 04:55:58.289629 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/515c81c7-10e3-4724-aafd-42431946c2c2-combined-ca-bundle\") pod \"515c81c7-10e3-4724-aafd-42431946c2c2\" (UID: \"515c81c7-10e3-4724-aafd-42431946c2c2\") " Nov 29 04:55:58 crc kubenswrapper[4799]: I1129 04:55:58.289714 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/515c81c7-10e3-4724-aafd-42431946c2c2-config-data\") pod \"515c81c7-10e3-4724-aafd-42431946c2c2\" (UID: \"515c81c7-10e3-4724-aafd-42431946c2c2\") " Nov 29 04:55:58 crc kubenswrapper[4799]: I1129 04:55:58.291496 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/515c81c7-10e3-4724-aafd-42431946c2c2-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "515c81c7-10e3-4724-aafd-42431946c2c2" (UID: "515c81c7-10e3-4724-aafd-42431946c2c2"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 04:55:58 crc kubenswrapper[4799]: I1129 04:55:58.297037 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/515c81c7-10e3-4724-aafd-42431946c2c2-kube-api-access-shprr" (OuterVolumeSpecName: "kube-api-access-shprr") pod "515c81c7-10e3-4724-aafd-42431946c2c2" (UID: "515c81c7-10e3-4724-aafd-42431946c2c2"). InnerVolumeSpecName "kube-api-access-shprr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:55:58 crc kubenswrapper[4799]: I1129 04:55:58.302856 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/515c81c7-10e3-4724-aafd-42431946c2c2-scripts" (OuterVolumeSpecName: "scripts") pod "515c81c7-10e3-4724-aafd-42431946c2c2" (UID: "515c81c7-10e3-4724-aafd-42431946c2c2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:55:58 crc kubenswrapper[4799]: I1129 04:55:58.312935 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/515c81c7-10e3-4724-aafd-42431946c2c2-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "515c81c7-10e3-4724-aafd-42431946c2c2" (UID: "515c81c7-10e3-4724-aafd-42431946c2c2"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:55:58 crc kubenswrapper[4799]: I1129 04:55:58.326684 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/515c81c7-10e3-4724-aafd-42431946c2c2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "515c81c7-10e3-4724-aafd-42431946c2c2" (UID: "515c81c7-10e3-4724-aafd-42431946c2c2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:55:58 crc kubenswrapper[4799]: I1129 04:55:58.349679 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/515c81c7-10e3-4724-aafd-42431946c2c2-config-data" (OuterVolumeSpecName: "config-data") pod "515c81c7-10e3-4724-aafd-42431946c2c2" (UID: "515c81c7-10e3-4724-aafd-42431946c2c2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:55:58 crc kubenswrapper[4799]: I1129 04:55:58.392296 4799 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/515c81c7-10e3-4724-aafd-42431946c2c2-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:58 crc kubenswrapper[4799]: I1129 04:55:58.392585 4799 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/515c81c7-10e3-4724-aafd-42431946c2c2-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:58 crc kubenswrapper[4799]: I1129 04:55:58.392650 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/515c81c7-10e3-4724-aafd-42431946c2c2-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:58 crc kubenswrapper[4799]: I1129 04:55:58.392704 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-shprr\" (UniqueName: \"kubernetes.io/projected/515c81c7-10e3-4724-aafd-42431946c2c2-kube-api-access-shprr\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:58 crc kubenswrapper[4799]: I1129 04:55:58.392769 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/515c81c7-10e3-4724-aafd-42431946c2c2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:58 crc kubenswrapper[4799]: I1129 04:55:58.392842 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/515c81c7-10e3-4724-aafd-42431946c2c2-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:58 crc kubenswrapper[4799]: I1129 04:55:58.538102 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6bb684768f-x97jl" Nov 29 04:55:58 crc kubenswrapper[4799]: I1129 04:55:58.610212 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7987f74bbc-m7qr2"] Nov 29 04:55:58 crc kubenswrapper[4799]: I1129 04:55:58.610515 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7987f74bbc-m7qr2" podUID="b80b0f55-9641-465a-b40c-f90244e53218" containerName="dnsmasq-dns" containerID="cri-o://53fdab9dd5d6c0e0517ca438e2da7bb79dd9d0c61301ad55fbe44c10d61c1015" gracePeriod=10 Nov 29 04:55:58 crc kubenswrapper[4799]: I1129 04:55:58.820314 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de975367-f439-4788-8ddb-b9f8af482589","Type":"ContainerStarted","Data":"22eed2795fed04c2ad4738a5b7ddd275701000eb0b9c938ced931a138b12ab81"} Nov 29 04:55:58 crc kubenswrapper[4799]: I1129 04:55:58.821842 4799 generic.go:334] "Generic (PLEG): container finished" podID="b80b0f55-9641-465a-b40c-f90244e53218" containerID="53fdab9dd5d6c0e0517ca438e2da7bb79dd9d0c61301ad55fbe44c10d61c1015" exitCode=0 Nov 29 04:55:58 crc kubenswrapper[4799]: I1129 04:55:58.821886 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7987f74bbc-m7qr2" event={"ID":"b80b0f55-9641-465a-b40c-f90244e53218","Type":"ContainerDied","Data":"53fdab9dd5d6c0e0517ca438e2da7bb79dd9d0c61301ad55fbe44c10d61c1015"} Nov 29 04:55:58 crc kubenswrapper[4799]: I1129 04:55:58.822967 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-btd8j" event={"ID":"515c81c7-10e3-4724-aafd-42431946c2c2","Type":"ContainerDied","Data":"8d5dc91c375bc0e9b87529b6d8e195a562d55d9e08494ad48c856841aeeb21df"} Nov 29 04:55:58 crc kubenswrapper[4799]: I1129 04:55:58.822991 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8d5dc91c375bc0e9b87529b6d8e195a562d55d9e08494ad48c856841aeeb21df" Nov 29 04:55:58 crc kubenswrapper[4799]: I1129 04:55:58.823049 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-btd8j" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.053821 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 29 04:55:59 crc kubenswrapper[4799]: E1129 04:55:59.054285 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="515c81c7-10e3-4724-aafd-42431946c2c2" containerName="cinder-db-sync" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.054301 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="515c81c7-10e3-4724-aafd-42431946c2c2" containerName="cinder-db-sync" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.054462 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="515c81c7-10e3-4724-aafd-42431946c2c2" containerName="cinder-db-sync" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.055436 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.056400 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.073606 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.073911 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-hvxx6" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.074163 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.074899 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.119820 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d97fcdd8f-9dgkt"] Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.121615 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d97fcdd8f-9dgkt" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.124359 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a6bcb3b8-0e44-42a4-8580-d514b1fe55be-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"a6bcb3b8-0e44-42a4-8580-d514b1fe55be\") " pod="openstack/cinder-scheduler-0" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.124396 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4bjd\" (UniqueName: \"kubernetes.io/projected/a6bcb3b8-0e44-42a4-8580-d514b1fe55be-kube-api-access-b4bjd\") pod \"cinder-scheduler-0\" (UID: \"a6bcb3b8-0e44-42a4-8580-d514b1fe55be\") " pod="openstack/cinder-scheduler-0" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.124428 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6bcb3b8-0e44-42a4-8580-d514b1fe55be-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"a6bcb3b8-0e44-42a4-8580-d514b1fe55be\") " pod="openstack/cinder-scheduler-0" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.124454 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6bcb3b8-0e44-42a4-8580-d514b1fe55be-config-data\") pod \"cinder-scheduler-0\" (UID: \"a6bcb3b8-0e44-42a4-8580-d514b1fe55be\") " pod="openstack/cinder-scheduler-0" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.124483 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6bcb3b8-0e44-42a4-8580-d514b1fe55be-scripts\") pod \"cinder-scheduler-0\" (UID: \"a6bcb3b8-0e44-42a4-8580-d514b1fe55be\") " pod="openstack/cinder-scheduler-0" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.124535 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a6bcb3b8-0e44-42a4-8580-d514b1fe55be-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"a6bcb3b8-0e44-42a4-8580-d514b1fe55be\") " pod="openstack/cinder-scheduler-0" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.200498 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d97fcdd8f-9dgkt"] Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.228028 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a6bcb3b8-0e44-42a4-8580-d514b1fe55be-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"a6bcb3b8-0e44-42a4-8580-d514b1fe55be\") " pod="openstack/cinder-scheduler-0" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.228083 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4bjd\" (UniqueName: \"kubernetes.io/projected/a6bcb3b8-0e44-42a4-8580-d514b1fe55be-kube-api-access-b4bjd\") pod \"cinder-scheduler-0\" (UID: \"a6bcb3b8-0e44-42a4-8580-d514b1fe55be\") " pod="openstack/cinder-scheduler-0" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.228116 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6bcb3b8-0e44-42a4-8580-d514b1fe55be-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"a6bcb3b8-0e44-42a4-8580-d514b1fe55be\") " pod="openstack/cinder-scheduler-0" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.228142 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c419e681-9099-4422-b04e-8faccbedc127-dns-svc\") pod \"dnsmasq-dns-6d97fcdd8f-9dgkt\" (UID: \"c419e681-9099-4422-b04e-8faccbedc127\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-9dgkt" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.228165 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c419e681-9099-4422-b04e-8faccbedc127-ovsdbserver-nb\") pod \"dnsmasq-dns-6d97fcdd8f-9dgkt\" (UID: \"c419e681-9099-4422-b04e-8faccbedc127\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-9dgkt" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.228189 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6bcb3b8-0e44-42a4-8580-d514b1fe55be-config-data\") pod \"cinder-scheduler-0\" (UID: \"a6bcb3b8-0e44-42a4-8580-d514b1fe55be\") " pod="openstack/cinder-scheduler-0" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.228228 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6bcb3b8-0e44-42a4-8580-d514b1fe55be-scripts\") pod \"cinder-scheduler-0\" (UID: \"a6bcb3b8-0e44-42a4-8580-d514b1fe55be\") " pod="openstack/cinder-scheduler-0" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.228255 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c419e681-9099-4422-b04e-8faccbedc127-config\") pod \"dnsmasq-dns-6d97fcdd8f-9dgkt\" (UID: \"c419e681-9099-4422-b04e-8faccbedc127\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-9dgkt" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.228299 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a6bcb3b8-0e44-42a4-8580-d514b1fe55be-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"a6bcb3b8-0e44-42a4-8580-d514b1fe55be\") " pod="openstack/cinder-scheduler-0" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.228334 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c419e681-9099-4422-b04e-8faccbedc127-ovsdbserver-sb\") pod \"dnsmasq-dns-6d97fcdd8f-9dgkt\" (UID: \"c419e681-9099-4422-b04e-8faccbedc127\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-9dgkt" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.228364 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzs8q\" (UniqueName: \"kubernetes.io/projected/c419e681-9099-4422-b04e-8faccbedc127-kube-api-access-gzs8q\") pod \"dnsmasq-dns-6d97fcdd8f-9dgkt\" (UID: \"c419e681-9099-4422-b04e-8faccbedc127\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-9dgkt" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.232951 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a6bcb3b8-0e44-42a4-8580-d514b1fe55be-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"a6bcb3b8-0e44-42a4-8580-d514b1fe55be\") " pod="openstack/cinder-scheduler-0" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.237898 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a6bcb3b8-0e44-42a4-8580-d514b1fe55be-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"a6bcb3b8-0e44-42a4-8580-d514b1fe55be\") " pod="openstack/cinder-scheduler-0" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.242057 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6bcb3b8-0e44-42a4-8580-d514b1fe55be-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"a6bcb3b8-0e44-42a4-8580-d514b1fe55be\") " pod="openstack/cinder-scheduler-0" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.260573 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6bcb3b8-0e44-42a4-8580-d514b1fe55be-scripts\") pod \"cinder-scheduler-0\" (UID: \"a6bcb3b8-0e44-42a4-8580-d514b1fe55be\") " pod="openstack/cinder-scheduler-0" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.266213 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6bcb3b8-0e44-42a4-8580-d514b1fe55be-config-data\") pod \"cinder-scheduler-0\" (UID: \"a6bcb3b8-0e44-42a4-8580-d514b1fe55be\") " pod="openstack/cinder-scheduler-0" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.271299 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4bjd\" (UniqueName: \"kubernetes.io/projected/a6bcb3b8-0e44-42a4-8580-d514b1fe55be-kube-api-access-b4bjd\") pod \"cinder-scheduler-0\" (UID: \"a6bcb3b8-0e44-42a4-8580-d514b1fe55be\") " pod="openstack/cinder-scheduler-0" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.333252 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c419e681-9099-4422-b04e-8faccbedc127-ovsdbserver-sb\") pod \"dnsmasq-dns-6d97fcdd8f-9dgkt\" (UID: \"c419e681-9099-4422-b04e-8faccbedc127\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-9dgkt" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.333322 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzs8q\" (UniqueName: \"kubernetes.io/projected/c419e681-9099-4422-b04e-8faccbedc127-kube-api-access-gzs8q\") pod \"dnsmasq-dns-6d97fcdd8f-9dgkt\" (UID: \"c419e681-9099-4422-b04e-8faccbedc127\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-9dgkt" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.333379 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c419e681-9099-4422-b04e-8faccbedc127-dns-svc\") pod \"dnsmasq-dns-6d97fcdd8f-9dgkt\" (UID: \"c419e681-9099-4422-b04e-8faccbedc127\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-9dgkt" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.333409 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c419e681-9099-4422-b04e-8faccbedc127-ovsdbserver-nb\") pod \"dnsmasq-dns-6d97fcdd8f-9dgkt\" (UID: \"c419e681-9099-4422-b04e-8faccbedc127\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-9dgkt" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.333456 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c419e681-9099-4422-b04e-8faccbedc127-config\") pod \"dnsmasq-dns-6d97fcdd8f-9dgkt\" (UID: \"c419e681-9099-4422-b04e-8faccbedc127\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-9dgkt" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.334459 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c419e681-9099-4422-b04e-8faccbedc127-config\") pod \"dnsmasq-dns-6d97fcdd8f-9dgkt\" (UID: \"c419e681-9099-4422-b04e-8faccbedc127\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-9dgkt" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.335287 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c419e681-9099-4422-b04e-8faccbedc127-ovsdbserver-sb\") pod \"dnsmasq-dns-6d97fcdd8f-9dgkt\" (UID: \"c419e681-9099-4422-b04e-8faccbedc127\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-9dgkt" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.336205 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c419e681-9099-4422-b04e-8faccbedc127-dns-svc\") pod \"dnsmasq-dns-6d97fcdd8f-9dgkt\" (UID: \"c419e681-9099-4422-b04e-8faccbedc127\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-9dgkt" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.339423 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c419e681-9099-4422-b04e-8faccbedc127-ovsdbserver-nb\") pod \"dnsmasq-dns-6d97fcdd8f-9dgkt\" (UID: \"c419e681-9099-4422-b04e-8faccbedc127\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-9dgkt" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.364107 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzs8q\" (UniqueName: \"kubernetes.io/projected/c419e681-9099-4422-b04e-8faccbedc127-kube-api-access-gzs8q\") pod \"dnsmasq-dns-6d97fcdd8f-9dgkt\" (UID: \"c419e681-9099-4422-b04e-8faccbedc127\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-9dgkt" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.371343 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.373612 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.380321 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.412920 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.436497 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1edb0ac0-196f-43cc-95c1-9c99138bc515-config-data\") pod \"cinder-api-0\" (UID: \"1edb0ac0-196f-43cc-95c1-9c99138bc515\") " pod="openstack/cinder-api-0" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.436594 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1edb0ac0-196f-43cc-95c1-9c99138bc515-logs\") pod \"cinder-api-0\" (UID: \"1edb0ac0-196f-43cc-95c1-9c99138bc515\") " pod="openstack/cinder-api-0" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.436618 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1edb0ac0-196f-43cc-95c1-9c99138bc515-config-data-custom\") pod \"cinder-api-0\" (UID: \"1edb0ac0-196f-43cc-95c1-9c99138bc515\") " pod="openstack/cinder-api-0" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.436655 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1edb0ac0-196f-43cc-95c1-9c99138bc515-scripts\") pod \"cinder-api-0\" (UID: \"1edb0ac0-196f-43cc-95c1-9c99138bc515\") " pod="openstack/cinder-api-0" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.436765 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1edb0ac0-196f-43cc-95c1-9c99138bc515-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1edb0ac0-196f-43cc-95c1-9c99138bc515\") " pod="openstack/cinder-api-0" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.436827 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1edb0ac0-196f-43cc-95c1-9c99138bc515-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1edb0ac0-196f-43cc-95c1-9c99138bc515\") " pod="openstack/cinder-api-0" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.436847 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4gxs\" (UniqueName: \"kubernetes.io/projected/1edb0ac0-196f-43cc-95c1-9c99138bc515-kube-api-access-j4gxs\") pod \"cinder-api-0\" (UID: \"1edb0ac0-196f-43cc-95c1-9c99138bc515\") " pod="openstack/cinder-api-0" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.455419 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7987f74bbc-m7qr2" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.472437 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.489480 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d97fcdd8f-9dgkt" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.537839 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s6n5x\" (UniqueName: \"kubernetes.io/projected/b80b0f55-9641-465a-b40c-f90244e53218-kube-api-access-s6n5x\") pod \"b80b0f55-9641-465a-b40c-f90244e53218\" (UID: \"b80b0f55-9641-465a-b40c-f90244e53218\") " Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.537912 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b80b0f55-9641-465a-b40c-f90244e53218-ovsdbserver-nb\") pod \"b80b0f55-9641-465a-b40c-f90244e53218\" (UID: \"b80b0f55-9641-465a-b40c-f90244e53218\") " Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.538516 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b80b0f55-9641-465a-b40c-f90244e53218-config\") pod \"b80b0f55-9641-465a-b40c-f90244e53218\" (UID: \"b80b0f55-9641-465a-b40c-f90244e53218\") " Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.538690 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b80b0f55-9641-465a-b40c-f90244e53218-dns-svc\") pod \"b80b0f55-9641-465a-b40c-f90244e53218\" (UID: \"b80b0f55-9641-465a-b40c-f90244e53218\") " Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.538750 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b80b0f55-9641-465a-b40c-f90244e53218-ovsdbserver-sb\") pod \"b80b0f55-9641-465a-b40c-f90244e53218\" (UID: \"b80b0f55-9641-465a-b40c-f90244e53218\") " Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.539094 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1edb0ac0-196f-43cc-95c1-9c99138bc515-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1edb0ac0-196f-43cc-95c1-9c99138bc515\") " pod="openstack/cinder-api-0" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.539165 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1edb0ac0-196f-43cc-95c1-9c99138bc515-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1edb0ac0-196f-43cc-95c1-9c99138bc515\") " pod="openstack/cinder-api-0" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.539192 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4gxs\" (UniqueName: \"kubernetes.io/projected/1edb0ac0-196f-43cc-95c1-9c99138bc515-kube-api-access-j4gxs\") pod \"cinder-api-0\" (UID: \"1edb0ac0-196f-43cc-95c1-9c99138bc515\") " pod="openstack/cinder-api-0" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.539246 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1edb0ac0-196f-43cc-95c1-9c99138bc515-config-data\") pod \"cinder-api-0\" (UID: \"1edb0ac0-196f-43cc-95c1-9c99138bc515\") " pod="openstack/cinder-api-0" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.539311 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1edb0ac0-196f-43cc-95c1-9c99138bc515-logs\") pod \"cinder-api-0\" (UID: \"1edb0ac0-196f-43cc-95c1-9c99138bc515\") " pod="openstack/cinder-api-0" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.539335 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1edb0ac0-196f-43cc-95c1-9c99138bc515-config-data-custom\") pod \"cinder-api-0\" (UID: \"1edb0ac0-196f-43cc-95c1-9c99138bc515\") " pod="openstack/cinder-api-0" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.539679 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1edb0ac0-196f-43cc-95c1-9c99138bc515-scripts\") pod \"cinder-api-0\" (UID: \"1edb0ac0-196f-43cc-95c1-9c99138bc515\") " pod="openstack/cinder-api-0" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.542277 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1edb0ac0-196f-43cc-95c1-9c99138bc515-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1edb0ac0-196f-43cc-95c1-9c99138bc515\") " pod="openstack/cinder-api-0" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.544154 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1edb0ac0-196f-43cc-95c1-9c99138bc515-logs\") pod \"cinder-api-0\" (UID: \"1edb0ac0-196f-43cc-95c1-9c99138bc515\") " pod="openstack/cinder-api-0" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.551298 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b80b0f55-9641-465a-b40c-f90244e53218-kube-api-access-s6n5x" (OuterVolumeSpecName: "kube-api-access-s6n5x") pod "b80b0f55-9641-465a-b40c-f90244e53218" (UID: "b80b0f55-9641-465a-b40c-f90244e53218"). InnerVolumeSpecName "kube-api-access-s6n5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.560932 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1edb0ac0-196f-43cc-95c1-9c99138bc515-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1edb0ac0-196f-43cc-95c1-9c99138bc515\") " pod="openstack/cinder-api-0" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.561350 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1edb0ac0-196f-43cc-95c1-9c99138bc515-config-data\") pod \"cinder-api-0\" (UID: \"1edb0ac0-196f-43cc-95c1-9c99138bc515\") " pod="openstack/cinder-api-0" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.561688 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1edb0ac0-196f-43cc-95c1-9c99138bc515-config-data-custom\") pod \"cinder-api-0\" (UID: \"1edb0ac0-196f-43cc-95c1-9c99138bc515\") " pod="openstack/cinder-api-0" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.564619 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1edb0ac0-196f-43cc-95c1-9c99138bc515-scripts\") pod \"cinder-api-0\" (UID: \"1edb0ac0-196f-43cc-95c1-9c99138bc515\") " pod="openstack/cinder-api-0" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.592299 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4gxs\" (UniqueName: \"kubernetes.io/projected/1edb0ac0-196f-43cc-95c1-9c99138bc515-kube-api-access-j4gxs\") pod \"cinder-api-0\" (UID: \"1edb0ac0-196f-43cc-95c1-9c99138bc515\") " pod="openstack/cinder-api-0" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.645509 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s6n5x\" (UniqueName: \"kubernetes.io/projected/b80b0f55-9641-465a-b40c-f90244e53218-kube-api-access-s6n5x\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.686345 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b80b0f55-9641-465a-b40c-f90244e53218-config" (OuterVolumeSpecName: "config") pod "b80b0f55-9641-465a-b40c-f90244e53218" (UID: "b80b0f55-9641-465a-b40c-f90244e53218"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.707455 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b80b0f55-9641-465a-b40c-f90244e53218-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b80b0f55-9641-465a-b40c-f90244e53218" (UID: "b80b0f55-9641-465a-b40c-f90244e53218"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.718134 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b80b0f55-9641-465a-b40c-f90244e53218-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b80b0f55-9641-465a-b40c-f90244e53218" (UID: "b80b0f55-9641-465a-b40c-f90244e53218"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.725626 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b80b0f55-9641-465a-b40c-f90244e53218-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b80b0f55-9641-465a-b40c-f90244e53218" (UID: "b80b0f55-9641-465a-b40c-f90244e53218"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.761311 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b80b0f55-9641-465a-b40c-f90244e53218-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.761341 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b80b0f55-9641-465a-b40c-f90244e53218-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.761358 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b80b0f55-9641-465a-b40c-f90244e53218-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.761371 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b80b0f55-9641-465a-b40c-f90244e53218-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.774096 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.925277 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de975367-f439-4788-8ddb-b9f8af482589","Type":"ContainerStarted","Data":"19f4901b0a5fd72073558e75cacf3cbb37d1bf2c2f64f0d5fb7ffc20577f1519"} Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.926935 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.931080 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7987f74bbc-m7qr2" event={"ID":"b80b0f55-9641-465a-b40c-f90244e53218","Type":"ContainerDied","Data":"53a604240972478ff9052bfa352a9d253572828d489daca54c06e91ebb6b4e85"} Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.931130 4799 scope.go:117] "RemoveContainer" containerID="53fdab9dd5d6c0e0517ca438e2da7bb79dd9d0c61301ad55fbe44c10d61c1015" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.931175 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7987f74bbc-m7qr2" Nov 29 04:55:59 crc kubenswrapper[4799]: I1129 04:55:59.952516 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.7961527349999997 podStartE2EDuration="6.952485185s" podCreationTimestamp="2025-11-29 04:55:53 +0000 UTC" firstStartedPulling="2025-11-29 04:55:55.172707901 +0000 UTC m=+1030.815638311" lastFinishedPulling="2025-11-29 04:55:59.329040361 +0000 UTC m=+1034.971970761" observedRunningTime="2025-11-29 04:55:59.949863561 +0000 UTC m=+1035.592793981" watchObservedRunningTime="2025-11-29 04:55:59.952485185 +0000 UTC m=+1035.595415585" Nov 29 04:56:00 crc kubenswrapper[4799]: I1129 04:55:59.997600 4799 scope.go:117] "RemoveContainer" containerID="0d0bf418898e7dc12fb2eecab71f5d4ba5cc8ee5f55a9bbbaeef94e5766137ea" Nov 29 04:56:00 crc kubenswrapper[4799]: I1129 04:56:00.089708 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7987f74bbc-m7qr2"] Nov 29 04:56:00 crc kubenswrapper[4799]: I1129 04:56:00.098066 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7987f74bbc-m7qr2"] Nov 29 04:56:00 crc kubenswrapper[4799]: I1129 04:56:00.187924 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 29 04:56:00 crc kubenswrapper[4799]: I1129 04:56:00.330477 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d97fcdd8f-9dgkt"] Nov 29 04:56:00 crc kubenswrapper[4799]: I1129 04:56:00.613648 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 29 04:56:00 crc kubenswrapper[4799]: W1129 04:56:00.651165 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1edb0ac0_196f_43cc_95c1_9c99138bc515.slice/crio-1b6b1be1436bd984aab36ec193ec67ad761ee7ee59cdc6fc1071fa0865cd82e7 WatchSource:0}: Error finding container 1b6b1be1436bd984aab36ec193ec67ad761ee7ee59cdc6fc1071fa0865cd82e7: Status 404 returned error can't find the container with id 1b6b1be1436bd984aab36ec193ec67ad761ee7ee59cdc6fc1071fa0865cd82e7 Nov 29 04:56:00 crc kubenswrapper[4799]: I1129 04:56:00.667947 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b80b0f55-9641-465a-b40c-f90244e53218" path="/var/lib/kubelet/pods/b80b0f55-9641-465a-b40c-f90244e53218/volumes" Nov 29 04:56:00 crc kubenswrapper[4799]: I1129 04:56:00.973987 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a6bcb3b8-0e44-42a4-8580-d514b1fe55be","Type":"ContainerStarted","Data":"a1a3e1eed69638e867d39f52d64b6cae6c4d3c5349f9915301298210b7006a59"} Nov 29 04:56:00 crc kubenswrapper[4799]: I1129 04:56:00.976434 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1edb0ac0-196f-43cc-95c1-9c99138bc515","Type":"ContainerStarted","Data":"1b6b1be1436bd984aab36ec193ec67ad761ee7ee59cdc6fc1071fa0865cd82e7"} Nov 29 04:56:00 crc kubenswrapper[4799]: I1129 04:56:00.980076 4799 generic.go:334] "Generic (PLEG): container finished" podID="c419e681-9099-4422-b04e-8faccbedc127" containerID="a59906d06ff21939ff4517a38752eb41f1cd150b909901896ecad59ea285e18c" exitCode=0 Nov 29 04:56:00 crc kubenswrapper[4799]: I1129 04:56:00.980175 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d97fcdd8f-9dgkt" event={"ID":"c419e681-9099-4422-b04e-8faccbedc127","Type":"ContainerDied","Data":"a59906d06ff21939ff4517a38752eb41f1cd150b909901896ecad59ea285e18c"} Nov 29 04:56:00 crc kubenswrapper[4799]: I1129 04:56:00.980250 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d97fcdd8f-9dgkt" event={"ID":"c419e681-9099-4422-b04e-8faccbedc127","Type":"ContainerStarted","Data":"a503e1abc5f477580dbaaa3fab6e9bdca2000a937ddc7bb49c9988af8d365f93"} Nov 29 04:56:01 crc kubenswrapper[4799]: I1129 04:56:01.216077 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6944bdb4b-6bt2h" Nov 29 04:56:01 crc kubenswrapper[4799]: I1129 04:56:01.391221 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6944bdb4b-6bt2h" Nov 29 04:56:01 crc kubenswrapper[4799]: I1129 04:56:01.594620 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 29 04:56:02 crc kubenswrapper[4799]: I1129 04:56:02.013102 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1edb0ac0-196f-43cc-95c1-9c99138bc515","Type":"ContainerStarted","Data":"218825a03106d7046447629934c8ed4172c54a526bf1c0d82a7c66c40521f4b2"} Nov 29 04:56:02 crc kubenswrapper[4799]: I1129 04:56:02.029410 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d97fcdd8f-9dgkt" event={"ID":"c419e681-9099-4422-b04e-8faccbedc127","Type":"ContainerStarted","Data":"296552b3cd8ba3aa9aef20b4bc6845e18cb64ff6248d63a812dda6adcbf9457f"} Nov 29 04:56:02 crc kubenswrapper[4799]: I1129 04:56:02.051278 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6d97fcdd8f-9dgkt" podStartSLOduration=3.051226697 podStartE2EDuration="3.051226697s" podCreationTimestamp="2025-11-29 04:55:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:56:02.050811086 +0000 UTC m=+1037.693741506" watchObservedRunningTime="2025-11-29 04:56:02.051226697 +0000 UTC m=+1037.694157107" Nov 29 04:56:03 crc kubenswrapper[4799]: I1129 04:56:03.039656 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1edb0ac0-196f-43cc-95c1-9c99138bc515","Type":"ContainerStarted","Data":"d9c222ce70688d5110b0c4ca06f9b1d4a3bf8f5da676a72f4d23573e23ecb5fb"} Nov 29 04:56:03 crc kubenswrapper[4799]: I1129 04:56:03.040173 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="1edb0ac0-196f-43cc-95c1-9c99138bc515" containerName="cinder-api-log" containerID="cri-o://218825a03106d7046447629934c8ed4172c54a526bf1c0d82a7c66c40521f4b2" gracePeriod=30 Nov 29 04:56:03 crc kubenswrapper[4799]: I1129 04:56:03.040439 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="1edb0ac0-196f-43cc-95c1-9c99138bc515" containerName="cinder-api" containerID="cri-o://d9c222ce70688d5110b0c4ca06f9b1d4a3bf8f5da676a72f4d23573e23ecb5fb" gracePeriod=30 Nov 29 04:56:03 crc kubenswrapper[4799]: I1129 04:56:03.040645 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 29 04:56:03 crc kubenswrapper[4799]: I1129 04:56:03.043570 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a6bcb3b8-0e44-42a4-8580-d514b1fe55be","Type":"ContainerStarted","Data":"39d467661ef7b11f70068a5bc51a8e5fb55c320cc9e6e483d33a066c636d038c"} Nov 29 04:56:03 crc kubenswrapper[4799]: I1129 04:56:03.043609 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d97fcdd8f-9dgkt" Nov 29 04:56:03 crc kubenswrapper[4799]: I1129 04:56:03.065447 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.065426034 podStartE2EDuration="4.065426034s" podCreationTimestamp="2025-11-29 04:55:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:56:03.061473758 +0000 UTC m=+1038.704404158" watchObservedRunningTime="2025-11-29 04:56:03.065426034 +0000 UTC m=+1038.708356434" Nov 29 04:56:03 crc kubenswrapper[4799]: I1129 04:56:03.735101 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 29 04:56:03 crc kubenswrapper[4799]: I1129 04:56:03.757853 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-fdfbddb56-lhh8r" Nov 29 04:56:03 crc kubenswrapper[4799]: I1129 04:56:03.823467 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j4gxs\" (UniqueName: \"kubernetes.io/projected/1edb0ac0-196f-43cc-95c1-9c99138bc515-kube-api-access-j4gxs\") pod \"1edb0ac0-196f-43cc-95c1-9c99138bc515\" (UID: \"1edb0ac0-196f-43cc-95c1-9c99138bc515\") " Nov 29 04:56:03 crc kubenswrapper[4799]: I1129 04:56:03.823902 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1edb0ac0-196f-43cc-95c1-9c99138bc515-config-data\") pod \"1edb0ac0-196f-43cc-95c1-9c99138bc515\" (UID: \"1edb0ac0-196f-43cc-95c1-9c99138bc515\") " Nov 29 04:56:03 crc kubenswrapper[4799]: I1129 04:56:03.823975 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1edb0ac0-196f-43cc-95c1-9c99138bc515-scripts\") pod \"1edb0ac0-196f-43cc-95c1-9c99138bc515\" (UID: \"1edb0ac0-196f-43cc-95c1-9c99138bc515\") " Nov 29 04:56:03 crc kubenswrapper[4799]: I1129 04:56:03.824054 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1edb0ac0-196f-43cc-95c1-9c99138bc515-config-data-custom\") pod \"1edb0ac0-196f-43cc-95c1-9c99138bc515\" (UID: \"1edb0ac0-196f-43cc-95c1-9c99138bc515\") " Nov 29 04:56:03 crc kubenswrapper[4799]: I1129 04:56:03.824151 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1edb0ac0-196f-43cc-95c1-9c99138bc515-logs\") pod \"1edb0ac0-196f-43cc-95c1-9c99138bc515\" (UID: \"1edb0ac0-196f-43cc-95c1-9c99138bc515\") " Nov 29 04:56:03 crc kubenswrapper[4799]: I1129 04:56:03.824186 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1edb0ac0-196f-43cc-95c1-9c99138bc515-etc-machine-id\") pod \"1edb0ac0-196f-43cc-95c1-9c99138bc515\" (UID: \"1edb0ac0-196f-43cc-95c1-9c99138bc515\") " Nov 29 04:56:03 crc kubenswrapper[4799]: I1129 04:56:03.824218 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1edb0ac0-196f-43cc-95c1-9c99138bc515-combined-ca-bundle\") pod \"1edb0ac0-196f-43cc-95c1-9c99138bc515\" (UID: \"1edb0ac0-196f-43cc-95c1-9c99138bc515\") " Nov 29 04:56:03 crc kubenswrapper[4799]: I1129 04:56:03.830007 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1edb0ac0-196f-43cc-95c1-9c99138bc515-kube-api-access-j4gxs" (OuterVolumeSpecName: "kube-api-access-j4gxs") pod "1edb0ac0-196f-43cc-95c1-9c99138bc515" (UID: "1edb0ac0-196f-43cc-95c1-9c99138bc515"). InnerVolumeSpecName "kube-api-access-j4gxs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:56:03 crc kubenswrapper[4799]: I1129 04:56:03.832154 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1edb0ac0-196f-43cc-95c1-9c99138bc515-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "1edb0ac0-196f-43cc-95c1-9c99138bc515" (UID: "1edb0ac0-196f-43cc-95c1-9c99138bc515"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 04:56:03 crc kubenswrapper[4799]: I1129 04:56:03.832364 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1edb0ac0-196f-43cc-95c1-9c99138bc515-logs" (OuterVolumeSpecName: "logs") pod "1edb0ac0-196f-43cc-95c1-9c99138bc515" (UID: "1edb0ac0-196f-43cc-95c1-9c99138bc515"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:56:03 crc kubenswrapper[4799]: I1129 04:56:03.833859 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1edb0ac0-196f-43cc-95c1-9c99138bc515-scripts" (OuterVolumeSpecName: "scripts") pod "1edb0ac0-196f-43cc-95c1-9c99138bc515" (UID: "1edb0ac0-196f-43cc-95c1-9c99138bc515"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:56:03 crc kubenswrapper[4799]: I1129 04:56:03.849096 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1edb0ac0-196f-43cc-95c1-9c99138bc515-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "1edb0ac0-196f-43cc-95c1-9c99138bc515" (UID: "1edb0ac0-196f-43cc-95c1-9c99138bc515"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:56:03 crc kubenswrapper[4799]: I1129 04:56:03.894663 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1edb0ac0-196f-43cc-95c1-9c99138bc515-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1edb0ac0-196f-43cc-95c1-9c99138bc515" (UID: "1edb0ac0-196f-43cc-95c1-9c99138bc515"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:56:03 crc kubenswrapper[4799]: I1129 04:56:03.917031 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1edb0ac0-196f-43cc-95c1-9c99138bc515-config-data" (OuterVolumeSpecName: "config-data") pod "1edb0ac0-196f-43cc-95c1-9c99138bc515" (UID: "1edb0ac0-196f-43cc-95c1-9c99138bc515"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:56:03 crc kubenswrapper[4799]: I1129 04:56:03.926649 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1edb0ac0-196f-43cc-95c1-9c99138bc515-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:03 crc kubenswrapper[4799]: I1129 04:56:03.927003 4799 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1edb0ac0-196f-43cc-95c1-9c99138bc515-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:03 crc kubenswrapper[4799]: I1129 04:56:03.927133 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1edb0ac0-196f-43cc-95c1-9c99138bc515-logs\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:03 crc kubenswrapper[4799]: I1129 04:56:03.927221 4799 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1edb0ac0-196f-43cc-95c1-9c99138bc515-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:03 crc kubenswrapper[4799]: I1129 04:56:03.927296 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1edb0ac0-196f-43cc-95c1-9c99138bc515-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:03 crc kubenswrapper[4799]: I1129 04:56:03.927378 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j4gxs\" (UniqueName: \"kubernetes.io/projected/1edb0ac0-196f-43cc-95c1-9c99138bc515-kube-api-access-j4gxs\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:03 crc kubenswrapper[4799]: I1129 04:56:03.927455 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1edb0ac0-196f-43cc-95c1-9c99138bc515-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.003777 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-fdfbddb56-lhh8r" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.100573 4799 generic.go:334] "Generic (PLEG): container finished" podID="1edb0ac0-196f-43cc-95c1-9c99138bc515" containerID="d9c222ce70688d5110b0c4ca06f9b1d4a3bf8f5da676a72f4d23573e23ecb5fb" exitCode=0 Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.100608 4799 generic.go:334] "Generic (PLEG): container finished" podID="1edb0ac0-196f-43cc-95c1-9c99138bc515" containerID="218825a03106d7046447629934c8ed4172c54a526bf1c0d82a7c66c40521f4b2" exitCode=143 Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.100650 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1edb0ac0-196f-43cc-95c1-9c99138bc515","Type":"ContainerDied","Data":"d9c222ce70688d5110b0c4ca06f9b1d4a3bf8f5da676a72f4d23573e23ecb5fb"} Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.100683 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1edb0ac0-196f-43cc-95c1-9c99138bc515","Type":"ContainerDied","Data":"218825a03106d7046447629934c8ed4172c54a526bf1c0d82a7c66c40521f4b2"} Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.100693 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1edb0ac0-196f-43cc-95c1-9c99138bc515","Type":"ContainerDied","Data":"1b6b1be1436bd984aab36ec193ec67ad761ee7ee59cdc6fc1071fa0865cd82e7"} Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.100711 4799 scope.go:117] "RemoveContainer" containerID="d9c222ce70688d5110b0c4ca06f9b1d4a3bf8f5da676a72f4d23573e23ecb5fb" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.101497 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.102860 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6944bdb4b-6bt2h"] Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.103168 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6944bdb4b-6bt2h" podUID="e964f0f8-6843-4a3c-917b-69372833af7a" containerName="barbican-api-log" containerID="cri-o://f073508318cdb64c0905a420f0d8f81d6ed3345066cadbf338a682fe615b86a9" gracePeriod=30 Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.103240 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6944bdb4b-6bt2h" podUID="e964f0f8-6843-4a3c-917b-69372833af7a" containerName="barbican-api" containerID="cri-o://8c6707e699584a656f44a5d76623c1304572bf57d4ffd926e1a45018d4470029" gracePeriod=30 Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.115482 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-6944bdb4b-6bt2h" podUID="e964f0f8-6843-4a3c-917b-69372833af7a" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.145:9311/healthcheck\": EOF" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.120141 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a6bcb3b8-0e44-42a4-8580-d514b1fe55be","Type":"ContainerStarted","Data":"b0da8e63661caa340f8fce1516d181419f503ef8a5fd40da84e549a3ab2fd7e9"} Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.153570 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=5.062481382 podStartE2EDuration="6.153552496s" podCreationTimestamp="2025-11-29 04:55:58 +0000 UTC" firstStartedPulling="2025-11-29 04:56:00.187201562 +0000 UTC m=+1035.830131962" lastFinishedPulling="2025-11-29 04:56:01.278272676 +0000 UTC m=+1036.921203076" observedRunningTime="2025-11-29 04:56:04.151431685 +0000 UTC m=+1039.794362085" watchObservedRunningTime="2025-11-29 04:56:04.153552496 +0000 UTC m=+1039.796482896" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.188978 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.207260 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.217358 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 29 04:56:04 crc kubenswrapper[4799]: E1129 04:56:04.217817 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b80b0f55-9641-465a-b40c-f90244e53218" containerName="dnsmasq-dns" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.217836 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="b80b0f55-9641-465a-b40c-f90244e53218" containerName="dnsmasq-dns" Nov 29 04:56:04 crc kubenswrapper[4799]: E1129 04:56:04.217853 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1edb0ac0-196f-43cc-95c1-9c99138bc515" containerName="cinder-api-log" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.217860 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="1edb0ac0-196f-43cc-95c1-9c99138bc515" containerName="cinder-api-log" Nov 29 04:56:04 crc kubenswrapper[4799]: E1129 04:56:04.217893 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b80b0f55-9641-465a-b40c-f90244e53218" containerName="init" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.217899 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="b80b0f55-9641-465a-b40c-f90244e53218" containerName="init" Nov 29 04:56:04 crc kubenswrapper[4799]: E1129 04:56:04.217909 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1edb0ac0-196f-43cc-95c1-9c99138bc515" containerName="cinder-api" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.217915 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="1edb0ac0-196f-43cc-95c1-9c99138bc515" containerName="cinder-api" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.218076 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="1edb0ac0-196f-43cc-95c1-9c99138bc515" containerName="cinder-api" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.218098 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="b80b0f55-9641-465a-b40c-f90244e53218" containerName="dnsmasq-dns" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.218110 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="1edb0ac0-196f-43cc-95c1-9c99138bc515" containerName="cinder-api-log" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.219383 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.224836 4799 scope.go:117] "RemoveContainer" containerID="218825a03106d7046447629934c8ed4172c54a526bf1c0d82a7c66c40521f4b2" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.225241 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.225553 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.225698 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.242130 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.283312 4799 scope.go:117] "RemoveContainer" containerID="d9c222ce70688d5110b0c4ca06f9b1d4a3bf8f5da676a72f4d23573e23ecb5fb" Nov 29 04:56:04 crc kubenswrapper[4799]: E1129 04:56:04.283876 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9c222ce70688d5110b0c4ca06f9b1d4a3bf8f5da676a72f4d23573e23ecb5fb\": container with ID starting with d9c222ce70688d5110b0c4ca06f9b1d4a3bf8f5da676a72f4d23573e23ecb5fb not found: ID does not exist" containerID="d9c222ce70688d5110b0c4ca06f9b1d4a3bf8f5da676a72f4d23573e23ecb5fb" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.283922 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9c222ce70688d5110b0c4ca06f9b1d4a3bf8f5da676a72f4d23573e23ecb5fb"} err="failed to get container status \"d9c222ce70688d5110b0c4ca06f9b1d4a3bf8f5da676a72f4d23573e23ecb5fb\": rpc error: code = NotFound desc = could not find container \"d9c222ce70688d5110b0c4ca06f9b1d4a3bf8f5da676a72f4d23573e23ecb5fb\": container with ID starting with d9c222ce70688d5110b0c4ca06f9b1d4a3bf8f5da676a72f4d23573e23ecb5fb not found: ID does not exist" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.283955 4799 scope.go:117] "RemoveContainer" containerID="218825a03106d7046447629934c8ed4172c54a526bf1c0d82a7c66c40521f4b2" Nov 29 04:56:04 crc kubenswrapper[4799]: E1129 04:56:04.284366 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"218825a03106d7046447629934c8ed4172c54a526bf1c0d82a7c66c40521f4b2\": container with ID starting with 218825a03106d7046447629934c8ed4172c54a526bf1c0d82a7c66c40521f4b2 not found: ID does not exist" containerID="218825a03106d7046447629934c8ed4172c54a526bf1c0d82a7c66c40521f4b2" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.284396 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"218825a03106d7046447629934c8ed4172c54a526bf1c0d82a7c66c40521f4b2"} err="failed to get container status \"218825a03106d7046447629934c8ed4172c54a526bf1c0d82a7c66c40521f4b2\": rpc error: code = NotFound desc = could not find container \"218825a03106d7046447629934c8ed4172c54a526bf1c0d82a7c66c40521f4b2\": container with ID starting with 218825a03106d7046447629934c8ed4172c54a526bf1c0d82a7c66c40521f4b2 not found: ID does not exist" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.284410 4799 scope.go:117] "RemoveContainer" containerID="d9c222ce70688d5110b0c4ca06f9b1d4a3bf8f5da676a72f4d23573e23ecb5fb" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.284758 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9c222ce70688d5110b0c4ca06f9b1d4a3bf8f5da676a72f4d23573e23ecb5fb"} err="failed to get container status \"d9c222ce70688d5110b0c4ca06f9b1d4a3bf8f5da676a72f4d23573e23ecb5fb\": rpc error: code = NotFound desc = could not find container \"d9c222ce70688d5110b0c4ca06f9b1d4a3bf8f5da676a72f4d23573e23ecb5fb\": container with ID starting with d9c222ce70688d5110b0c4ca06f9b1d4a3bf8f5da676a72f4d23573e23ecb5fb not found: ID does not exist" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.284782 4799 scope.go:117] "RemoveContainer" containerID="218825a03106d7046447629934c8ed4172c54a526bf1c0d82a7c66c40521f4b2" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.284995 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"218825a03106d7046447629934c8ed4172c54a526bf1c0d82a7c66c40521f4b2"} err="failed to get container status \"218825a03106d7046447629934c8ed4172c54a526bf1c0d82a7c66c40521f4b2\": rpc error: code = NotFound desc = could not find container \"218825a03106d7046447629934c8ed4172c54a526bf1c0d82a7c66c40521f4b2\": container with ID starting with 218825a03106d7046447629934c8ed4172c54a526bf1c0d82a7c66c40521f4b2 not found: ID does not exist" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.346633 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/64042cd5-9bc5-4371-b063-753befa911c2-logs\") pod \"cinder-api-0\" (UID: \"64042cd5-9bc5-4371-b063-753befa911c2\") " pod="openstack/cinder-api-0" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.346706 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9r9z\" (UniqueName: \"kubernetes.io/projected/64042cd5-9bc5-4371-b063-753befa911c2-kube-api-access-h9r9z\") pod \"cinder-api-0\" (UID: \"64042cd5-9bc5-4371-b063-753befa911c2\") " pod="openstack/cinder-api-0" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.346754 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/64042cd5-9bc5-4371-b063-753befa911c2-config-data-custom\") pod \"cinder-api-0\" (UID: \"64042cd5-9bc5-4371-b063-753befa911c2\") " pod="openstack/cinder-api-0" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.346832 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/64042cd5-9bc5-4371-b063-753befa911c2-scripts\") pod \"cinder-api-0\" (UID: \"64042cd5-9bc5-4371-b063-753befa911c2\") " pod="openstack/cinder-api-0" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.346867 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64042cd5-9bc5-4371-b063-753befa911c2-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"64042cd5-9bc5-4371-b063-753befa911c2\") " pod="openstack/cinder-api-0" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.346905 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64042cd5-9bc5-4371-b063-753befa911c2-config-data\") pod \"cinder-api-0\" (UID: \"64042cd5-9bc5-4371-b063-753befa911c2\") " pod="openstack/cinder-api-0" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.346925 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/64042cd5-9bc5-4371-b063-753befa911c2-public-tls-certs\") pod \"cinder-api-0\" (UID: \"64042cd5-9bc5-4371-b063-753befa911c2\") " pod="openstack/cinder-api-0" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.346963 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/64042cd5-9bc5-4371-b063-753befa911c2-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"64042cd5-9bc5-4371-b063-753befa911c2\") " pod="openstack/cinder-api-0" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.346995 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/64042cd5-9bc5-4371-b063-753befa911c2-etc-machine-id\") pod \"cinder-api-0\" (UID: \"64042cd5-9bc5-4371-b063-753befa911c2\") " pod="openstack/cinder-api-0" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.448956 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/64042cd5-9bc5-4371-b063-753befa911c2-scripts\") pod \"cinder-api-0\" (UID: \"64042cd5-9bc5-4371-b063-753befa911c2\") " pod="openstack/cinder-api-0" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.449017 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64042cd5-9bc5-4371-b063-753befa911c2-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"64042cd5-9bc5-4371-b063-753befa911c2\") " pod="openstack/cinder-api-0" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.449042 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64042cd5-9bc5-4371-b063-753befa911c2-config-data\") pod \"cinder-api-0\" (UID: \"64042cd5-9bc5-4371-b063-753befa911c2\") " pod="openstack/cinder-api-0" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.449061 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/64042cd5-9bc5-4371-b063-753befa911c2-public-tls-certs\") pod \"cinder-api-0\" (UID: \"64042cd5-9bc5-4371-b063-753befa911c2\") " pod="openstack/cinder-api-0" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.449114 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/64042cd5-9bc5-4371-b063-753befa911c2-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"64042cd5-9bc5-4371-b063-753befa911c2\") " pod="openstack/cinder-api-0" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.449139 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/64042cd5-9bc5-4371-b063-753befa911c2-etc-machine-id\") pod \"cinder-api-0\" (UID: \"64042cd5-9bc5-4371-b063-753befa911c2\") " pod="openstack/cinder-api-0" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.449189 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/64042cd5-9bc5-4371-b063-753befa911c2-logs\") pod \"cinder-api-0\" (UID: \"64042cd5-9bc5-4371-b063-753befa911c2\") " pod="openstack/cinder-api-0" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.449225 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9r9z\" (UniqueName: \"kubernetes.io/projected/64042cd5-9bc5-4371-b063-753befa911c2-kube-api-access-h9r9z\") pod \"cinder-api-0\" (UID: \"64042cd5-9bc5-4371-b063-753befa911c2\") " pod="openstack/cinder-api-0" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.449257 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/64042cd5-9bc5-4371-b063-753befa911c2-config-data-custom\") pod \"cinder-api-0\" (UID: \"64042cd5-9bc5-4371-b063-753befa911c2\") " pod="openstack/cinder-api-0" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.452154 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/64042cd5-9bc5-4371-b063-753befa911c2-logs\") pod \"cinder-api-0\" (UID: \"64042cd5-9bc5-4371-b063-753befa911c2\") " pod="openstack/cinder-api-0" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.452392 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/64042cd5-9bc5-4371-b063-753befa911c2-etc-machine-id\") pod \"cinder-api-0\" (UID: \"64042cd5-9bc5-4371-b063-753befa911c2\") " pod="openstack/cinder-api-0" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.459761 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64042cd5-9bc5-4371-b063-753befa911c2-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"64042cd5-9bc5-4371-b063-753befa911c2\") " pod="openstack/cinder-api-0" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.459811 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/64042cd5-9bc5-4371-b063-753befa911c2-config-data-custom\") pod \"cinder-api-0\" (UID: \"64042cd5-9bc5-4371-b063-753befa911c2\") " pod="openstack/cinder-api-0" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.460171 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/64042cd5-9bc5-4371-b063-753befa911c2-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"64042cd5-9bc5-4371-b063-753befa911c2\") " pod="openstack/cinder-api-0" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.460385 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64042cd5-9bc5-4371-b063-753befa911c2-config-data\") pod \"cinder-api-0\" (UID: \"64042cd5-9bc5-4371-b063-753befa911c2\") " pod="openstack/cinder-api-0" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.460705 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/64042cd5-9bc5-4371-b063-753befa911c2-scripts\") pod \"cinder-api-0\" (UID: \"64042cd5-9bc5-4371-b063-753befa911c2\") " pod="openstack/cinder-api-0" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.463201 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/64042cd5-9bc5-4371-b063-753befa911c2-public-tls-certs\") pod \"cinder-api-0\" (UID: \"64042cd5-9bc5-4371-b063-753befa911c2\") " pod="openstack/cinder-api-0" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.472375 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9r9z\" (UniqueName: \"kubernetes.io/projected/64042cd5-9bc5-4371-b063-753befa911c2-kube-api-access-h9r9z\") pod \"cinder-api-0\" (UID: \"64042cd5-9bc5-4371-b063-753befa911c2\") " pod="openstack/cinder-api-0" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.473565 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.593083 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 29 04:56:04 crc kubenswrapper[4799]: I1129 04:56:04.674022 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1edb0ac0-196f-43cc-95c1-9c99138bc515" path="/var/lib/kubelet/pods/1edb0ac0-196f-43cc-95c1-9c99138bc515/volumes" Nov 29 04:56:05 crc kubenswrapper[4799]: I1129 04:56:05.093100 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 29 04:56:05 crc kubenswrapper[4799]: I1129 04:56:05.134345 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"64042cd5-9bc5-4371-b063-753befa911c2","Type":"ContainerStarted","Data":"46ced3efc2dc1901d75408ae3d96404f4dbf4e656ef1159ff47245cba891f485"} Nov 29 04:56:05 crc kubenswrapper[4799]: I1129 04:56:05.136719 4799 generic.go:334] "Generic (PLEG): container finished" podID="e964f0f8-6843-4a3c-917b-69372833af7a" containerID="f073508318cdb64c0905a420f0d8f81d6ed3345066cadbf338a682fe615b86a9" exitCode=143 Nov 29 04:56:05 crc kubenswrapper[4799]: I1129 04:56:05.137818 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6944bdb4b-6bt2h" event={"ID":"e964f0f8-6843-4a3c-917b-69372833af7a","Type":"ContainerDied","Data":"f073508318cdb64c0905a420f0d8f81d6ed3345066cadbf338a682fe615b86a9"} Nov 29 04:56:06 crc kubenswrapper[4799]: I1129 04:56:06.067009 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-dc694ffdd-sm2s5" Nov 29 04:56:06 crc kubenswrapper[4799]: I1129 04:56:06.071976 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-dc694ffdd-sm2s5" Nov 29 04:56:06 crc kubenswrapper[4799]: I1129 04:56:06.163534 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"64042cd5-9bc5-4371-b063-753befa911c2","Type":"ContainerStarted","Data":"dd7e3d0a4485db536c1c0aed6c00245683623c8db2cd255725cbf2a4ca697193"} Nov 29 04:56:07 crc kubenswrapper[4799]: I1129 04:56:07.178624 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"64042cd5-9bc5-4371-b063-753befa911c2","Type":"ContainerStarted","Data":"58d004567da4001a62a56721c5d60c66ba6bdbb05582b46fd3a89976b08d52f2"} Nov 29 04:56:07 crc kubenswrapper[4799]: I1129 04:56:07.179008 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 29 04:56:07 crc kubenswrapper[4799]: I1129 04:56:07.210316 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.210288985 podStartE2EDuration="3.210288985s" podCreationTimestamp="2025-11-29 04:56:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:56:07.204476763 +0000 UTC m=+1042.847407163" watchObservedRunningTime="2025-11-29 04:56:07.210288985 +0000 UTC m=+1042.853219385" Nov 29 04:56:08 crc kubenswrapper[4799]: I1129 04:56:08.371444 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-757bb6ddd6-9wg88" Nov 29 04:56:09 crc kubenswrapper[4799]: I1129 04:56:09.493097 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6d97fcdd8f-9dgkt" Nov 29 04:56:09 crc kubenswrapper[4799]: I1129 04:56:09.561232 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bb684768f-x97jl"] Nov 29 04:56:09 crc kubenswrapper[4799]: I1129 04:56:09.561581 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6bb684768f-x97jl" podUID="0ae506e6-d401-4fa6-b4ce-8db525ef694e" containerName="dnsmasq-dns" containerID="cri-o://98920f3f0ac9e5a6e35fcc7c3647d9f155b455fe2b3f70acf845c665ed352fb3" gracePeriod=10 Nov 29 04:56:09 crc kubenswrapper[4799]: I1129 04:56:09.599351 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6944bdb4b-6bt2h" podUID="e964f0f8-6843-4a3c-917b-69372833af7a" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.145:9311/healthcheck\": read tcp 10.217.0.2:45900->10.217.0.145:9311: read: connection reset by peer" Nov 29 04:56:09 crc kubenswrapper[4799]: I1129 04:56:09.599464 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6944bdb4b-6bt2h" podUID="e964f0f8-6843-4a3c-917b-69372833af7a" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.145:9311/healthcheck\": read tcp 10.217.0.2:45894->10.217.0.145:9311: read: connection reset by peer" Nov 29 04:56:09 crc kubenswrapper[4799]: I1129 04:56:09.877574 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 29 04:56:09 crc kubenswrapper[4799]: I1129 04:56:09.934622 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.135858 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb684768f-x97jl" Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.224088 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ae506e6-d401-4fa6-b4ce-8db525ef694e-config\") pod \"0ae506e6-d401-4fa6-b4ce-8db525ef694e\" (UID: \"0ae506e6-d401-4fa6-b4ce-8db525ef694e\") " Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.224188 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0ae506e6-d401-4fa6-b4ce-8db525ef694e-ovsdbserver-nb\") pod \"0ae506e6-d401-4fa6-b4ce-8db525ef694e\" (UID: \"0ae506e6-d401-4fa6-b4ce-8db525ef694e\") " Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.224209 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0ae506e6-d401-4fa6-b4ce-8db525ef694e-ovsdbserver-sb\") pod \"0ae506e6-d401-4fa6-b4ce-8db525ef694e\" (UID: \"0ae506e6-d401-4fa6-b4ce-8db525ef694e\") " Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.224239 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2f5wf\" (UniqueName: \"kubernetes.io/projected/0ae506e6-d401-4fa6-b4ce-8db525ef694e-kube-api-access-2f5wf\") pod \"0ae506e6-d401-4fa6-b4ce-8db525ef694e\" (UID: \"0ae506e6-d401-4fa6-b4ce-8db525ef694e\") " Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.224317 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0ae506e6-d401-4fa6-b4ce-8db525ef694e-dns-svc\") pod \"0ae506e6-d401-4fa6-b4ce-8db525ef694e\" (UID: \"0ae506e6-d401-4fa6-b4ce-8db525ef694e\") " Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.227900 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6944bdb4b-6bt2h" Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.241799 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ae506e6-d401-4fa6-b4ce-8db525ef694e-kube-api-access-2f5wf" (OuterVolumeSpecName: "kube-api-access-2f5wf") pod "0ae506e6-d401-4fa6-b4ce-8db525ef694e" (UID: "0ae506e6-d401-4fa6-b4ce-8db525ef694e"). InnerVolumeSpecName "kube-api-access-2f5wf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.252070 4799 generic.go:334] "Generic (PLEG): container finished" podID="0ae506e6-d401-4fa6-b4ce-8db525ef694e" containerID="98920f3f0ac9e5a6e35fcc7c3647d9f155b455fe2b3f70acf845c665ed352fb3" exitCode=0 Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.252153 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb684768f-x97jl" event={"ID":"0ae506e6-d401-4fa6-b4ce-8db525ef694e","Type":"ContainerDied","Data":"98920f3f0ac9e5a6e35fcc7c3647d9f155b455fe2b3f70acf845c665ed352fb3"} Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.252189 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb684768f-x97jl" event={"ID":"0ae506e6-d401-4fa6-b4ce-8db525ef694e","Type":"ContainerDied","Data":"c5f4c51a4869b03f3a4338ee5573abf6799fbc1e6c086a3f1aec8b9426b57642"} Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.252206 4799 scope.go:117] "RemoveContainer" containerID="98920f3f0ac9e5a6e35fcc7c3647d9f155b455fe2b3f70acf845c665ed352fb3" Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.252344 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb684768f-x97jl" Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.270561 4799 generic.go:334] "Generic (PLEG): container finished" podID="e964f0f8-6843-4a3c-917b-69372833af7a" containerID="8c6707e699584a656f44a5d76623c1304572bf57d4ffd926e1a45018d4470029" exitCode=0 Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.270645 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6944bdb4b-6bt2h" event={"ID":"e964f0f8-6843-4a3c-917b-69372833af7a","Type":"ContainerDied","Data":"8c6707e699584a656f44a5d76623c1304572bf57d4ffd926e1a45018d4470029"} Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.270713 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6944bdb4b-6bt2h" event={"ID":"e964f0f8-6843-4a3c-917b-69372833af7a","Type":"ContainerDied","Data":"c001f01269ba72b4d3b08214fc658088bb0920bdb8a9a18f6ea9c52cf5028e9d"} Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.270728 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6944bdb4b-6bt2h" Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.270855 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="a6bcb3b8-0e44-42a4-8580-d514b1fe55be" containerName="cinder-scheduler" containerID="cri-o://39d467661ef7b11f70068a5bc51a8e5fb55c320cc9e6e483d33a066c636d038c" gracePeriod=30 Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.271771 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="a6bcb3b8-0e44-42a4-8580-d514b1fe55be" containerName="probe" containerID="cri-o://b0da8e63661caa340f8fce1516d181419f503ef8a5fd40da84e549a3ab2fd7e9" gracePeriod=30 Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.292076 4799 scope.go:117] "RemoveContainer" containerID="d9fa614974d8b4dda6d470d2fddb9c45c4a2825e1d075b1fe65142f6749caba5" Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.313286 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ae506e6-d401-4fa6-b4ce-8db525ef694e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0ae506e6-d401-4fa6-b4ce-8db525ef694e" (UID: "0ae506e6-d401-4fa6-b4ce-8db525ef694e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.313556 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ae506e6-d401-4fa6-b4ce-8db525ef694e-config" (OuterVolumeSpecName: "config") pod "0ae506e6-d401-4fa6-b4ce-8db525ef694e" (UID: "0ae506e6-d401-4fa6-b4ce-8db525ef694e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.318591 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ae506e6-d401-4fa6-b4ce-8db525ef694e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0ae506e6-d401-4fa6-b4ce-8db525ef694e" (UID: "0ae506e6-d401-4fa6-b4ce-8db525ef694e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.325067 4799 scope.go:117] "RemoveContainer" containerID="98920f3f0ac9e5a6e35fcc7c3647d9f155b455fe2b3f70acf845c665ed352fb3" Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.325869 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e964f0f8-6843-4a3c-917b-69372833af7a-logs\") pod \"e964f0f8-6843-4a3c-917b-69372833af7a\" (UID: \"e964f0f8-6843-4a3c-917b-69372833af7a\") " Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.325944 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e964f0f8-6843-4a3c-917b-69372833af7a-config-data-custom\") pod \"e964f0f8-6843-4a3c-917b-69372833af7a\" (UID: \"e964f0f8-6843-4a3c-917b-69372833af7a\") " Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.326030 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e964f0f8-6843-4a3c-917b-69372833af7a-combined-ca-bundle\") pod \"e964f0f8-6843-4a3c-917b-69372833af7a\" (UID: \"e964f0f8-6843-4a3c-917b-69372833af7a\") " Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.326084 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zngmq\" (UniqueName: \"kubernetes.io/projected/e964f0f8-6843-4a3c-917b-69372833af7a-kube-api-access-zngmq\") pod \"e964f0f8-6843-4a3c-917b-69372833af7a\" (UID: \"e964f0f8-6843-4a3c-917b-69372833af7a\") " Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.326192 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e964f0f8-6843-4a3c-917b-69372833af7a-config-data\") pod \"e964f0f8-6843-4a3c-917b-69372833af7a\" (UID: \"e964f0f8-6843-4a3c-917b-69372833af7a\") " Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.326699 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e964f0f8-6843-4a3c-917b-69372833af7a-logs" (OuterVolumeSpecName: "logs") pod "e964f0f8-6843-4a3c-917b-69372833af7a" (UID: "e964f0f8-6843-4a3c-917b-69372833af7a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.326751 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ae506e6-d401-4fa6-b4ce-8db525ef694e-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.326765 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0ae506e6-d401-4fa6-b4ce-8db525ef694e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.326778 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2f5wf\" (UniqueName: \"kubernetes.io/projected/0ae506e6-d401-4fa6-b4ce-8db525ef694e-kube-api-access-2f5wf\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.326802 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0ae506e6-d401-4fa6-b4ce-8db525ef694e-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:10 crc kubenswrapper[4799]: E1129 04:56:10.327416 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98920f3f0ac9e5a6e35fcc7c3647d9f155b455fe2b3f70acf845c665ed352fb3\": container with ID starting with 98920f3f0ac9e5a6e35fcc7c3647d9f155b455fe2b3f70acf845c665ed352fb3 not found: ID does not exist" containerID="98920f3f0ac9e5a6e35fcc7c3647d9f155b455fe2b3f70acf845c665ed352fb3" Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.327467 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98920f3f0ac9e5a6e35fcc7c3647d9f155b455fe2b3f70acf845c665ed352fb3"} err="failed to get container status \"98920f3f0ac9e5a6e35fcc7c3647d9f155b455fe2b3f70acf845c665ed352fb3\": rpc error: code = NotFound desc = could not find container \"98920f3f0ac9e5a6e35fcc7c3647d9f155b455fe2b3f70acf845c665ed352fb3\": container with ID starting with 98920f3f0ac9e5a6e35fcc7c3647d9f155b455fe2b3f70acf845c665ed352fb3 not found: ID does not exist" Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.327506 4799 scope.go:117] "RemoveContainer" containerID="d9fa614974d8b4dda6d470d2fddb9c45c4a2825e1d075b1fe65142f6749caba5" Nov 29 04:56:10 crc kubenswrapper[4799]: E1129 04:56:10.328730 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9fa614974d8b4dda6d470d2fddb9c45c4a2825e1d075b1fe65142f6749caba5\": container with ID starting with d9fa614974d8b4dda6d470d2fddb9c45c4a2825e1d075b1fe65142f6749caba5 not found: ID does not exist" containerID="d9fa614974d8b4dda6d470d2fddb9c45c4a2825e1d075b1fe65142f6749caba5" Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.328809 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9fa614974d8b4dda6d470d2fddb9c45c4a2825e1d075b1fe65142f6749caba5"} err="failed to get container status \"d9fa614974d8b4dda6d470d2fddb9c45c4a2825e1d075b1fe65142f6749caba5\": rpc error: code = NotFound desc = could not find container \"d9fa614974d8b4dda6d470d2fddb9c45c4a2825e1d075b1fe65142f6749caba5\": container with ID starting with d9fa614974d8b4dda6d470d2fddb9c45c4a2825e1d075b1fe65142f6749caba5 not found: ID does not exist" Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.328852 4799 scope.go:117] "RemoveContainer" containerID="8c6707e699584a656f44a5d76623c1304572bf57d4ffd926e1a45018d4470029" Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.331059 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e964f0f8-6843-4a3c-917b-69372833af7a-kube-api-access-zngmq" (OuterVolumeSpecName: "kube-api-access-zngmq") pod "e964f0f8-6843-4a3c-917b-69372833af7a" (UID: "e964f0f8-6843-4a3c-917b-69372833af7a"). InnerVolumeSpecName "kube-api-access-zngmq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.331095 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ae506e6-d401-4fa6-b4ce-8db525ef694e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0ae506e6-d401-4fa6-b4ce-8db525ef694e" (UID: "0ae506e6-d401-4fa6-b4ce-8db525ef694e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.331688 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e964f0f8-6843-4a3c-917b-69372833af7a-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e964f0f8-6843-4a3c-917b-69372833af7a" (UID: "e964f0f8-6843-4a3c-917b-69372833af7a"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.351292 4799 scope.go:117] "RemoveContainer" containerID="f073508318cdb64c0905a420f0d8f81d6ed3345066cadbf338a682fe615b86a9" Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.353536 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e964f0f8-6843-4a3c-917b-69372833af7a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e964f0f8-6843-4a3c-917b-69372833af7a" (UID: "e964f0f8-6843-4a3c-917b-69372833af7a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.383497 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e964f0f8-6843-4a3c-917b-69372833af7a-config-data" (OuterVolumeSpecName: "config-data") pod "e964f0f8-6843-4a3c-917b-69372833af7a" (UID: "e964f0f8-6843-4a3c-917b-69372833af7a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.386093 4799 scope.go:117] "RemoveContainer" containerID="8c6707e699584a656f44a5d76623c1304572bf57d4ffd926e1a45018d4470029" Nov 29 04:56:10 crc kubenswrapper[4799]: E1129 04:56:10.386708 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c6707e699584a656f44a5d76623c1304572bf57d4ffd926e1a45018d4470029\": container with ID starting with 8c6707e699584a656f44a5d76623c1304572bf57d4ffd926e1a45018d4470029 not found: ID does not exist" containerID="8c6707e699584a656f44a5d76623c1304572bf57d4ffd926e1a45018d4470029" Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.386758 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c6707e699584a656f44a5d76623c1304572bf57d4ffd926e1a45018d4470029"} err="failed to get container status \"8c6707e699584a656f44a5d76623c1304572bf57d4ffd926e1a45018d4470029\": rpc error: code = NotFound desc = could not find container \"8c6707e699584a656f44a5d76623c1304572bf57d4ffd926e1a45018d4470029\": container with ID starting with 8c6707e699584a656f44a5d76623c1304572bf57d4ffd926e1a45018d4470029 not found: ID does not exist" Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.386809 4799 scope.go:117] "RemoveContainer" containerID="f073508318cdb64c0905a420f0d8f81d6ed3345066cadbf338a682fe615b86a9" Nov 29 04:56:10 crc kubenswrapper[4799]: E1129 04:56:10.387548 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f073508318cdb64c0905a420f0d8f81d6ed3345066cadbf338a682fe615b86a9\": container with ID starting with f073508318cdb64c0905a420f0d8f81d6ed3345066cadbf338a682fe615b86a9 not found: ID does not exist" containerID="f073508318cdb64c0905a420f0d8f81d6ed3345066cadbf338a682fe615b86a9" Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.387573 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f073508318cdb64c0905a420f0d8f81d6ed3345066cadbf338a682fe615b86a9"} err="failed to get container status \"f073508318cdb64c0905a420f0d8f81d6ed3345066cadbf338a682fe615b86a9\": rpc error: code = NotFound desc = could not find container \"f073508318cdb64c0905a420f0d8f81d6ed3345066cadbf338a682fe615b86a9\": container with ID starting with f073508318cdb64c0905a420f0d8f81d6ed3345066cadbf338a682fe615b86a9 not found: ID does not exist" Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.428861 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zngmq\" (UniqueName: \"kubernetes.io/projected/e964f0f8-6843-4a3c-917b-69372833af7a-kube-api-access-zngmq\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.428898 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e964f0f8-6843-4a3c-917b-69372833af7a-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.428911 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0ae506e6-d401-4fa6-b4ce-8db525ef694e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.428921 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e964f0f8-6843-4a3c-917b-69372833af7a-logs\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.428931 4799 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e964f0f8-6843-4a3c-917b-69372833af7a-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.428943 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e964f0f8-6843-4a3c-917b-69372833af7a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.591357 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bb684768f-x97jl"] Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.606684 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bb684768f-x97jl"] Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.619246 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6944bdb4b-6bt2h"] Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.626392 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-6944bdb4b-6bt2h"] Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.670745 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ae506e6-d401-4fa6-b4ce-8db525ef694e" path="/var/lib/kubelet/pods/0ae506e6-d401-4fa6-b4ce-8db525ef694e/volumes" Nov 29 04:56:10 crc kubenswrapper[4799]: I1129 04:56:10.671549 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e964f0f8-6843-4a3c-917b-69372833af7a" path="/var/lib/kubelet/pods/e964f0f8-6843-4a3c-917b-69372833af7a/volumes" Nov 29 04:56:11 crc kubenswrapper[4799]: I1129 04:56:11.092368 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-d557d86d8-6zk7x" Nov 29 04:56:11 crc kubenswrapper[4799]: I1129 04:56:11.288626 4799 generic.go:334] "Generic (PLEG): container finished" podID="a6bcb3b8-0e44-42a4-8580-d514b1fe55be" containerID="b0da8e63661caa340f8fce1516d181419f503ef8a5fd40da84e549a3ab2fd7e9" exitCode=0 Nov 29 04:56:11 crc kubenswrapper[4799]: I1129 04:56:11.288719 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a6bcb3b8-0e44-42a4-8580-d514b1fe55be","Type":"ContainerDied","Data":"b0da8e63661caa340f8fce1516d181419f503ef8a5fd40da84e549a3ab2fd7e9"} Nov 29 04:56:11 crc kubenswrapper[4799]: I1129 04:56:11.544333 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 29 04:56:11 crc kubenswrapper[4799]: E1129 04:56:11.544865 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e964f0f8-6843-4a3c-917b-69372833af7a" containerName="barbican-api" Nov 29 04:56:11 crc kubenswrapper[4799]: I1129 04:56:11.544886 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e964f0f8-6843-4a3c-917b-69372833af7a" containerName="barbican-api" Nov 29 04:56:11 crc kubenswrapper[4799]: E1129 04:56:11.544907 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ae506e6-d401-4fa6-b4ce-8db525ef694e" containerName="init" Nov 29 04:56:11 crc kubenswrapper[4799]: I1129 04:56:11.544914 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ae506e6-d401-4fa6-b4ce-8db525ef694e" containerName="init" Nov 29 04:56:11 crc kubenswrapper[4799]: E1129 04:56:11.544935 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e964f0f8-6843-4a3c-917b-69372833af7a" containerName="barbican-api-log" Nov 29 04:56:11 crc kubenswrapper[4799]: I1129 04:56:11.544942 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e964f0f8-6843-4a3c-917b-69372833af7a" containerName="barbican-api-log" Nov 29 04:56:11 crc kubenswrapper[4799]: E1129 04:56:11.544968 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ae506e6-d401-4fa6-b4ce-8db525ef694e" containerName="dnsmasq-dns" Nov 29 04:56:11 crc kubenswrapper[4799]: I1129 04:56:11.544981 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ae506e6-d401-4fa6-b4ce-8db525ef694e" containerName="dnsmasq-dns" Nov 29 04:56:11 crc kubenswrapper[4799]: I1129 04:56:11.545173 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="e964f0f8-6843-4a3c-917b-69372833af7a" containerName="barbican-api-log" Nov 29 04:56:11 crc kubenswrapper[4799]: I1129 04:56:11.545192 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ae506e6-d401-4fa6-b4ce-8db525ef694e" containerName="dnsmasq-dns" Nov 29 04:56:11 crc kubenswrapper[4799]: I1129 04:56:11.545205 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="e964f0f8-6843-4a3c-917b-69372833af7a" containerName="barbican-api" Nov 29 04:56:11 crc kubenswrapper[4799]: I1129 04:56:11.545981 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 29 04:56:11 crc kubenswrapper[4799]: I1129 04:56:11.548711 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Nov 29 04:56:11 crc kubenswrapper[4799]: I1129 04:56:11.548759 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Nov 29 04:56:11 crc kubenswrapper[4799]: I1129 04:56:11.550623 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-x96pl" Nov 29 04:56:11 crc kubenswrapper[4799]: I1129 04:56:11.556112 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 29 04:56:11 crc kubenswrapper[4799]: I1129 04:56:11.652818 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/548ff572-b2d0-45ec-a131-0ba87044b4ff-openstack-config-secret\") pod \"openstackclient\" (UID: \"548ff572-b2d0-45ec-a131-0ba87044b4ff\") " pod="openstack/openstackclient" Nov 29 04:56:11 crc kubenswrapper[4799]: I1129 04:56:11.652864 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97gp8\" (UniqueName: \"kubernetes.io/projected/548ff572-b2d0-45ec-a131-0ba87044b4ff-kube-api-access-97gp8\") pod \"openstackclient\" (UID: \"548ff572-b2d0-45ec-a131-0ba87044b4ff\") " pod="openstack/openstackclient" Nov 29 04:56:11 crc kubenswrapper[4799]: I1129 04:56:11.652912 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/548ff572-b2d0-45ec-a131-0ba87044b4ff-combined-ca-bundle\") pod \"openstackclient\" (UID: \"548ff572-b2d0-45ec-a131-0ba87044b4ff\") " pod="openstack/openstackclient" Nov 29 04:56:11 crc kubenswrapper[4799]: I1129 04:56:11.652947 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/548ff572-b2d0-45ec-a131-0ba87044b4ff-openstack-config\") pod \"openstackclient\" (UID: \"548ff572-b2d0-45ec-a131-0ba87044b4ff\") " pod="openstack/openstackclient" Nov 29 04:56:11 crc kubenswrapper[4799]: I1129 04:56:11.754623 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/548ff572-b2d0-45ec-a131-0ba87044b4ff-openstack-config-secret\") pod \"openstackclient\" (UID: \"548ff572-b2d0-45ec-a131-0ba87044b4ff\") " pod="openstack/openstackclient" Nov 29 04:56:11 crc kubenswrapper[4799]: I1129 04:56:11.754680 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97gp8\" (UniqueName: \"kubernetes.io/projected/548ff572-b2d0-45ec-a131-0ba87044b4ff-kube-api-access-97gp8\") pod \"openstackclient\" (UID: \"548ff572-b2d0-45ec-a131-0ba87044b4ff\") " pod="openstack/openstackclient" Nov 29 04:56:11 crc kubenswrapper[4799]: I1129 04:56:11.754728 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/548ff572-b2d0-45ec-a131-0ba87044b4ff-combined-ca-bundle\") pod \"openstackclient\" (UID: \"548ff572-b2d0-45ec-a131-0ba87044b4ff\") " pod="openstack/openstackclient" Nov 29 04:56:11 crc kubenswrapper[4799]: I1129 04:56:11.754762 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/548ff572-b2d0-45ec-a131-0ba87044b4ff-openstack-config\") pod \"openstackclient\" (UID: \"548ff572-b2d0-45ec-a131-0ba87044b4ff\") " pod="openstack/openstackclient" Nov 29 04:56:11 crc kubenswrapper[4799]: I1129 04:56:11.755773 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/548ff572-b2d0-45ec-a131-0ba87044b4ff-openstack-config\") pod \"openstackclient\" (UID: \"548ff572-b2d0-45ec-a131-0ba87044b4ff\") " pod="openstack/openstackclient" Nov 29 04:56:11 crc kubenswrapper[4799]: I1129 04:56:11.759802 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/548ff572-b2d0-45ec-a131-0ba87044b4ff-openstack-config-secret\") pod \"openstackclient\" (UID: \"548ff572-b2d0-45ec-a131-0ba87044b4ff\") " pod="openstack/openstackclient" Nov 29 04:56:11 crc kubenswrapper[4799]: I1129 04:56:11.772281 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/548ff572-b2d0-45ec-a131-0ba87044b4ff-combined-ca-bundle\") pod \"openstackclient\" (UID: \"548ff572-b2d0-45ec-a131-0ba87044b4ff\") " pod="openstack/openstackclient" Nov 29 04:56:11 crc kubenswrapper[4799]: I1129 04:56:11.772491 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97gp8\" (UniqueName: \"kubernetes.io/projected/548ff572-b2d0-45ec-a131-0ba87044b4ff-kube-api-access-97gp8\") pod \"openstackclient\" (UID: \"548ff572-b2d0-45ec-a131-0ba87044b4ff\") " pod="openstack/openstackclient" Nov 29 04:56:11 crc kubenswrapper[4799]: I1129 04:56:11.870718 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 29 04:56:12 crc kubenswrapper[4799]: I1129 04:56:12.364224 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 29 04:56:12 crc kubenswrapper[4799]: W1129 04:56:12.369045 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod548ff572_b2d0_45ec_a131_0ba87044b4ff.slice/crio-f92fb28bd61d09c5955979383c8df8f9d0fb800bcd0843751db453a699d859e1 WatchSource:0}: Error finding container f92fb28bd61d09c5955979383c8df8f9d0fb800bcd0843751db453a699d859e1: Status 404 returned error can't find the container with id f92fb28bd61d09c5955979383c8df8f9d0fb800bcd0843751db453a699d859e1 Nov 29 04:56:13 crc kubenswrapper[4799]: I1129 04:56:13.317116 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"548ff572-b2d0-45ec-a131-0ba87044b4ff","Type":"ContainerStarted","Data":"f92fb28bd61d09c5955979383c8df8f9d0fb800bcd0843751db453a699d859e1"} Nov 29 04:56:13 crc kubenswrapper[4799]: I1129 04:56:13.382515 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-669f887b5-74klb" Nov 29 04:56:13 crc kubenswrapper[4799]: I1129 04:56:13.454655 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-d557d86d8-6zk7x"] Nov 29 04:56:13 crc kubenswrapper[4799]: I1129 04:56:13.455031 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-d557d86d8-6zk7x" podUID="f21cec8f-8cd4-4ecd-9800-3bb0c482fb66" containerName="neutron-api" containerID="cri-o://0206e122fb25f53bced3ad79d205461cb91bdecdbdf24990b6664f0037824b4c" gracePeriod=30 Nov 29 04:56:13 crc kubenswrapper[4799]: I1129 04:56:13.455185 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-d557d86d8-6zk7x" podUID="f21cec8f-8cd4-4ecd-9800-3bb0c482fb66" containerName="neutron-httpd" containerID="cri-o://af15be66d22796a789def34857a97e75d27b749e255595a4c9a405ba71d3db05" gracePeriod=30 Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.181249 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.305099 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6bcb3b8-0e44-42a4-8580-d514b1fe55be-config-data\") pod \"a6bcb3b8-0e44-42a4-8580-d514b1fe55be\" (UID: \"a6bcb3b8-0e44-42a4-8580-d514b1fe55be\") " Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.305402 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6bcb3b8-0e44-42a4-8580-d514b1fe55be-combined-ca-bundle\") pod \"a6bcb3b8-0e44-42a4-8580-d514b1fe55be\" (UID: \"a6bcb3b8-0e44-42a4-8580-d514b1fe55be\") " Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.305521 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6bcb3b8-0e44-42a4-8580-d514b1fe55be-scripts\") pod \"a6bcb3b8-0e44-42a4-8580-d514b1fe55be\" (UID: \"a6bcb3b8-0e44-42a4-8580-d514b1fe55be\") " Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.305593 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a6bcb3b8-0e44-42a4-8580-d514b1fe55be-etc-machine-id\") pod \"a6bcb3b8-0e44-42a4-8580-d514b1fe55be\" (UID: \"a6bcb3b8-0e44-42a4-8580-d514b1fe55be\") " Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.305625 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4bjd\" (UniqueName: \"kubernetes.io/projected/a6bcb3b8-0e44-42a4-8580-d514b1fe55be-kube-api-access-b4bjd\") pod \"a6bcb3b8-0e44-42a4-8580-d514b1fe55be\" (UID: \"a6bcb3b8-0e44-42a4-8580-d514b1fe55be\") " Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.305684 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a6bcb3b8-0e44-42a4-8580-d514b1fe55be-config-data-custom\") pod \"a6bcb3b8-0e44-42a4-8580-d514b1fe55be\" (UID: \"a6bcb3b8-0e44-42a4-8580-d514b1fe55be\") " Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.305694 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a6bcb3b8-0e44-42a4-8580-d514b1fe55be-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "a6bcb3b8-0e44-42a4-8580-d514b1fe55be" (UID: "a6bcb3b8-0e44-42a4-8580-d514b1fe55be"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.306093 4799 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a6bcb3b8-0e44-42a4-8580-d514b1fe55be-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.315579 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6bcb3b8-0e44-42a4-8580-d514b1fe55be-scripts" (OuterVolumeSpecName: "scripts") pod "a6bcb3b8-0e44-42a4-8580-d514b1fe55be" (UID: "a6bcb3b8-0e44-42a4-8580-d514b1fe55be"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.322058 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6bcb3b8-0e44-42a4-8580-d514b1fe55be-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a6bcb3b8-0e44-42a4-8580-d514b1fe55be" (UID: "a6bcb3b8-0e44-42a4-8580-d514b1fe55be"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.330868 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6bcb3b8-0e44-42a4-8580-d514b1fe55be-kube-api-access-b4bjd" (OuterVolumeSpecName: "kube-api-access-b4bjd") pod "a6bcb3b8-0e44-42a4-8580-d514b1fe55be" (UID: "a6bcb3b8-0e44-42a4-8580-d514b1fe55be"). InnerVolumeSpecName "kube-api-access-b4bjd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.334070 4799 generic.go:334] "Generic (PLEG): container finished" podID="a6bcb3b8-0e44-42a4-8580-d514b1fe55be" containerID="39d467661ef7b11f70068a5bc51a8e5fb55c320cc9e6e483d33a066c636d038c" exitCode=0 Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.334168 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a6bcb3b8-0e44-42a4-8580-d514b1fe55be","Type":"ContainerDied","Data":"39d467661ef7b11f70068a5bc51a8e5fb55c320cc9e6e483d33a066c636d038c"} Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.334211 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a6bcb3b8-0e44-42a4-8580-d514b1fe55be","Type":"ContainerDied","Data":"a1a3e1eed69638e867d39f52d64b6cae6c4d3c5349f9915301298210b7006a59"} Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.334235 4799 scope.go:117] "RemoveContainer" containerID="b0da8e63661caa340f8fce1516d181419f503ef8a5fd40da84e549a3ab2fd7e9" Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.334440 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.343906 4799 generic.go:334] "Generic (PLEG): container finished" podID="f21cec8f-8cd4-4ecd-9800-3bb0c482fb66" containerID="af15be66d22796a789def34857a97e75d27b749e255595a4c9a405ba71d3db05" exitCode=0 Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.343962 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d557d86d8-6zk7x" event={"ID":"f21cec8f-8cd4-4ecd-9800-3bb0c482fb66","Type":"ContainerDied","Data":"af15be66d22796a789def34857a97e75d27b749e255595a4c9a405ba71d3db05"} Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.385370 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6bcb3b8-0e44-42a4-8580-d514b1fe55be-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a6bcb3b8-0e44-42a4-8580-d514b1fe55be" (UID: "a6bcb3b8-0e44-42a4-8580-d514b1fe55be"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.389474 4799 scope.go:117] "RemoveContainer" containerID="39d467661ef7b11f70068a5bc51a8e5fb55c320cc9e6e483d33a066c636d038c" Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.408235 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6bcb3b8-0e44-42a4-8580-d514b1fe55be-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.408277 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6bcb3b8-0e44-42a4-8580-d514b1fe55be-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.408291 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4bjd\" (UniqueName: \"kubernetes.io/projected/a6bcb3b8-0e44-42a4-8580-d514b1fe55be-kube-api-access-b4bjd\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.408305 4799 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a6bcb3b8-0e44-42a4-8580-d514b1fe55be-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.436457 4799 scope.go:117] "RemoveContainer" containerID="b0da8e63661caa340f8fce1516d181419f503ef8a5fd40da84e549a3ab2fd7e9" Nov 29 04:56:14 crc kubenswrapper[4799]: E1129 04:56:14.438286 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0da8e63661caa340f8fce1516d181419f503ef8a5fd40da84e549a3ab2fd7e9\": container with ID starting with b0da8e63661caa340f8fce1516d181419f503ef8a5fd40da84e549a3ab2fd7e9 not found: ID does not exist" containerID="b0da8e63661caa340f8fce1516d181419f503ef8a5fd40da84e549a3ab2fd7e9" Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.438338 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0da8e63661caa340f8fce1516d181419f503ef8a5fd40da84e549a3ab2fd7e9"} err="failed to get container status \"b0da8e63661caa340f8fce1516d181419f503ef8a5fd40da84e549a3ab2fd7e9\": rpc error: code = NotFound desc = could not find container \"b0da8e63661caa340f8fce1516d181419f503ef8a5fd40da84e549a3ab2fd7e9\": container with ID starting with b0da8e63661caa340f8fce1516d181419f503ef8a5fd40da84e549a3ab2fd7e9 not found: ID does not exist" Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.438369 4799 scope.go:117] "RemoveContainer" containerID="39d467661ef7b11f70068a5bc51a8e5fb55c320cc9e6e483d33a066c636d038c" Nov 29 04:56:14 crc kubenswrapper[4799]: E1129 04:56:14.439569 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39d467661ef7b11f70068a5bc51a8e5fb55c320cc9e6e483d33a066c636d038c\": container with ID starting with 39d467661ef7b11f70068a5bc51a8e5fb55c320cc9e6e483d33a066c636d038c not found: ID does not exist" containerID="39d467661ef7b11f70068a5bc51a8e5fb55c320cc9e6e483d33a066c636d038c" Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.439605 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39d467661ef7b11f70068a5bc51a8e5fb55c320cc9e6e483d33a066c636d038c"} err="failed to get container status \"39d467661ef7b11f70068a5bc51a8e5fb55c320cc9e6e483d33a066c636d038c\": rpc error: code = NotFound desc = could not find container \"39d467661ef7b11f70068a5bc51a8e5fb55c320cc9e6e483d33a066c636d038c\": container with ID starting with 39d467661ef7b11f70068a5bc51a8e5fb55c320cc9e6e483d33a066c636d038c not found: ID does not exist" Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.444474 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6bcb3b8-0e44-42a4-8580-d514b1fe55be-config-data" (OuterVolumeSpecName: "config-data") pod "a6bcb3b8-0e44-42a4-8580-d514b1fe55be" (UID: "a6bcb3b8-0e44-42a4-8580-d514b1fe55be"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.510412 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6bcb3b8-0e44-42a4-8580-d514b1fe55be-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.687552 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.789512 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.814103 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 29 04:56:14 crc kubenswrapper[4799]: E1129 04:56:14.816569 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6bcb3b8-0e44-42a4-8580-d514b1fe55be" containerName="probe" Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.816596 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6bcb3b8-0e44-42a4-8580-d514b1fe55be" containerName="probe" Nov 29 04:56:14 crc kubenswrapper[4799]: E1129 04:56:14.816640 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6bcb3b8-0e44-42a4-8580-d514b1fe55be" containerName="cinder-scheduler" Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.816647 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6bcb3b8-0e44-42a4-8580-d514b1fe55be" containerName="cinder-scheduler" Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.820375 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6bcb3b8-0e44-42a4-8580-d514b1fe55be" containerName="cinder-scheduler" Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.820496 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6bcb3b8-0e44-42a4-8580-d514b1fe55be" containerName="probe" Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.826873 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.830921 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.834383 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.922095 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c7456e1-235c-4bb3-a587-f7e3c620749e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6c7456e1-235c-4bb3-a587-f7e3c620749e\") " pod="openstack/cinder-scheduler-0" Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.922181 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c7456e1-235c-4bb3-a587-f7e3c620749e-scripts\") pod \"cinder-scheduler-0\" (UID: \"6c7456e1-235c-4bb3-a587-f7e3c620749e\") " pod="openstack/cinder-scheduler-0" Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.922214 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7p4m\" (UniqueName: \"kubernetes.io/projected/6c7456e1-235c-4bb3-a587-f7e3c620749e-kube-api-access-b7p4m\") pod \"cinder-scheduler-0\" (UID: \"6c7456e1-235c-4bb3-a587-f7e3c620749e\") " pod="openstack/cinder-scheduler-0" Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.922251 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6c7456e1-235c-4bb3-a587-f7e3c620749e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6c7456e1-235c-4bb3-a587-f7e3c620749e\") " pod="openstack/cinder-scheduler-0" Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.922285 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c7456e1-235c-4bb3-a587-f7e3c620749e-config-data\") pod \"cinder-scheduler-0\" (UID: \"6c7456e1-235c-4bb3-a587-f7e3c620749e\") " pod="openstack/cinder-scheduler-0" Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.922349 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6c7456e1-235c-4bb3-a587-f7e3c620749e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6c7456e1-235c-4bb3-a587-f7e3c620749e\") " pod="openstack/cinder-scheduler-0" Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.937759 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.939829 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="de975367-f439-4788-8ddb-b9f8af482589" containerName="ceilometer-central-agent" containerID="cri-o://59ea15e9044a8c7f1247d7fb32c1ad6ab35c795668363a83216a4a7e201564c0" gracePeriod=30 Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.940782 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="de975367-f439-4788-8ddb-b9f8af482589" containerName="sg-core" containerID="cri-o://22eed2795fed04c2ad4738a5b7ddd275701000eb0b9c938ced931a138b12ab81" gracePeriod=30 Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.940907 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="de975367-f439-4788-8ddb-b9f8af482589" containerName="proxy-httpd" containerID="cri-o://19f4901b0a5fd72073558e75cacf3cbb37d1bf2c2f64f0d5fb7ffc20577f1519" gracePeriod=30 Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.940900 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="de975367-f439-4788-8ddb-b9f8af482589" containerName="ceilometer-notification-agent" containerID="cri-o://fbd3059919aa4e1d1d2943fd052b2861ed1b731729d22d5383486cc878702a33" gracePeriod=30 Nov 29 04:56:14 crc kubenswrapper[4799]: I1129 04:56:14.948198 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="de975367-f439-4788-8ddb-b9f8af482589" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.147:3000/\": EOF" Nov 29 04:56:15 crc kubenswrapper[4799]: I1129 04:56:15.024332 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6c7456e1-235c-4bb3-a587-f7e3c620749e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6c7456e1-235c-4bb3-a587-f7e3c620749e\") " pod="openstack/cinder-scheduler-0" Nov 29 04:56:15 crc kubenswrapper[4799]: I1129 04:56:15.024414 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c7456e1-235c-4bb3-a587-f7e3c620749e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6c7456e1-235c-4bb3-a587-f7e3c620749e\") " pod="openstack/cinder-scheduler-0" Nov 29 04:56:15 crc kubenswrapper[4799]: I1129 04:56:15.024452 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6c7456e1-235c-4bb3-a587-f7e3c620749e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6c7456e1-235c-4bb3-a587-f7e3c620749e\") " pod="openstack/cinder-scheduler-0" Nov 29 04:56:15 crc kubenswrapper[4799]: I1129 04:56:15.024458 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c7456e1-235c-4bb3-a587-f7e3c620749e-scripts\") pod \"cinder-scheduler-0\" (UID: \"6c7456e1-235c-4bb3-a587-f7e3c620749e\") " pod="openstack/cinder-scheduler-0" Nov 29 04:56:15 crc kubenswrapper[4799]: I1129 04:56:15.024717 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7p4m\" (UniqueName: \"kubernetes.io/projected/6c7456e1-235c-4bb3-a587-f7e3c620749e-kube-api-access-b7p4m\") pod \"cinder-scheduler-0\" (UID: \"6c7456e1-235c-4bb3-a587-f7e3c620749e\") " pod="openstack/cinder-scheduler-0" Nov 29 04:56:15 crc kubenswrapper[4799]: I1129 04:56:15.025197 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6c7456e1-235c-4bb3-a587-f7e3c620749e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6c7456e1-235c-4bb3-a587-f7e3c620749e\") " pod="openstack/cinder-scheduler-0" Nov 29 04:56:15 crc kubenswrapper[4799]: I1129 04:56:15.025245 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c7456e1-235c-4bb3-a587-f7e3c620749e-config-data\") pod \"cinder-scheduler-0\" (UID: \"6c7456e1-235c-4bb3-a587-f7e3c620749e\") " pod="openstack/cinder-scheduler-0" Nov 29 04:56:15 crc kubenswrapper[4799]: I1129 04:56:15.029520 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c7456e1-235c-4bb3-a587-f7e3c620749e-scripts\") pod \"cinder-scheduler-0\" (UID: \"6c7456e1-235c-4bb3-a587-f7e3c620749e\") " pod="openstack/cinder-scheduler-0" Nov 29 04:56:15 crc kubenswrapper[4799]: I1129 04:56:15.029677 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c7456e1-235c-4bb3-a587-f7e3c620749e-config-data\") pod \"cinder-scheduler-0\" (UID: \"6c7456e1-235c-4bb3-a587-f7e3c620749e\") " pod="openstack/cinder-scheduler-0" Nov 29 04:56:15 crc kubenswrapper[4799]: I1129 04:56:15.035209 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c7456e1-235c-4bb3-a587-f7e3c620749e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6c7456e1-235c-4bb3-a587-f7e3c620749e\") " pod="openstack/cinder-scheduler-0" Nov 29 04:56:15 crc kubenswrapper[4799]: I1129 04:56:15.042877 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7p4m\" (UniqueName: \"kubernetes.io/projected/6c7456e1-235c-4bb3-a587-f7e3c620749e-kube-api-access-b7p4m\") pod \"cinder-scheduler-0\" (UID: \"6c7456e1-235c-4bb3-a587-f7e3c620749e\") " pod="openstack/cinder-scheduler-0" Nov 29 04:56:15 crc kubenswrapper[4799]: I1129 04:56:15.048125 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6c7456e1-235c-4bb3-a587-f7e3c620749e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6c7456e1-235c-4bb3-a587-f7e3c620749e\") " pod="openstack/cinder-scheduler-0" Nov 29 04:56:15 crc kubenswrapper[4799]: I1129 04:56:15.164906 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 29 04:56:15 crc kubenswrapper[4799]: I1129 04:56:15.359961 4799 generic.go:334] "Generic (PLEG): container finished" podID="de975367-f439-4788-8ddb-b9f8af482589" containerID="19f4901b0a5fd72073558e75cacf3cbb37d1bf2c2f64f0d5fb7ffc20577f1519" exitCode=0 Nov 29 04:56:15 crc kubenswrapper[4799]: I1129 04:56:15.360344 4799 generic.go:334] "Generic (PLEG): container finished" podID="de975367-f439-4788-8ddb-b9f8af482589" containerID="22eed2795fed04c2ad4738a5b7ddd275701000eb0b9c938ced931a138b12ab81" exitCode=2 Nov 29 04:56:15 crc kubenswrapper[4799]: I1129 04:56:15.360407 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de975367-f439-4788-8ddb-b9f8af482589","Type":"ContainerDied","Data":"19f4901b0a5fd72073558e75cacf3cbb37d1bf2c2f64f0d5fb7ffc20577f1519"} Nov 29 04:56:15 crc kubenswrapper[4799]: I1129 04:56:15.360448 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de975367-f439-4788-8ddb-b9f8af482589","Type":"ContainerDied","Data":"22eed2795fed04c2ad4738a5b7ddd275701000eb0b9c938ced931a138b12ab81"} Nov 29 04:56:15 crc kubenswrapper[4799]: I1129 04:56:15.686608 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 29 04:56:15 crc kubenswrapper[4799]: W1129 04:56:15.702437 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6c7456e1_235c_4bb3_a587_f7e3c620749e.slice/crio-bd23841ed1f690c53cc9164f62851851bfd3aa97d3788063146f59436354ee8a WatchSource:0}: Error finding container bd23841ed1f690c53cc9164f62851851bfd3aa97d3788063146f59436354ee8a: Status 404 returned error can't find the container with id bd23841ed1f690c53cc9164f62851851bfd3aa97d3788063146f59436354ee8a Nov 29 04:56:16 crc kubenswrapper[4799]: I1129 04:56:16.398231 4799 generic.go:334] "Generic (PLEG): container finished" podID="de975367-f439-4788-8ddb-b9f8af482589" containerID="59ea15e9044a8c7f1247d7fb32c1ad6ab35c795668363a83216a4a7e201564c0" exitCode=0 Nov 29 04:56:16 crc kubenswrapper[4799]: I1129 04:56:16.398569 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de975367-f439-4788-8ddb-b9f8af482589","Type":"ContainerDied","Data":"59ea15e9044a8c7f1247d7fb32c1ad6ab35c795668363a83216a4a7e201564c0"} Nov 29 04:56:16 crc kubenswrapper[4799]: I1129 04:56:16.400060 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6c7456e1-235c-4bb3-a587-f7e3c620749e","Type":"ContainerStarted","Data":"bd23841ed1f690c53cc9164f62851851bfd3aa97d3788063146f59436354ee8a"} Nov 29 04:56:16 crc kubenswrapper[4799]: I1129 04:56:16.674158 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6bcb3b8-0e44-42a4-8580-d514b1fe55be" path="/var/lib/kubelet/pods/a6bcb3b8-0e44-42a4-8580-d514b1fe55be/volumes" Nov 29 04:56:17 crc kubenswrapper[4799]: I1129 04:56:17.164862 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 29 04:56:17 crc kubenswrapper[4799]: I1129 04:56:17.416721 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6c7456e1-235c-4bb3-a587-f7e3c620749e","Type":"ContainerStarted","Data":"63e6693432046dfa7e7b5ad876fc22402b1f2524c92e61f76d86f892fbc2d2b8"} Nov 29 04:56:17 crc kubenswrapper[4799]: I1129 04:56:17.416773 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6c7456e1-235c-4bb3-a587-f7e3c620749e","Type":"ContainerStarted","Data":"2edd1b45ac8ff596b7d29a3dd6871f8be9afcceb059b051ac4f993c0b9e12e7d"} Nov 29 04:56:17 crc kubenswrapper[4799]: I1129 04:56:17.443383 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.443353707 podStartE2EDuration="3.443353707s" podCreationTimestamp="2025-11-29 04:56:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:56:17.44016661 +0000 UTC m=+1053.083097020" watchObservedRunningTime="2025-11-29 04:56:17.443353707 +0000 UTC m=+1053.086284107" Nov 29 04:56:20 crc kubenswrapper[4799]: I1129 04:56:20.165623 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 29 04:56:21 crc kubenswrapper[4799]: I1129 04:56:21.474361 4799 generic.go:334] "Generic (PLEG): container finished" podID="de975367-f439-4788-8ddb-b9f8af482589" containerID="fbd3059919aa4e1d1d2943fd052b2861ed1b731729d22d5383486cc878702a33" exitCode=0 Nov 29 04:56:21 crc kubenswrapper[4799]: I1129 04:56:21.474494 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de975367-f439-4788-8ddb-b9f8af482589","Type":"ContainerDied","Data":"fbd3059919aa4e1d1d2943fd052b2861ed1b731729d22d5383486cc878702a33"} Nov 29 04:56:23 crc kubenswrapper[4799]: I1129 04:56:23.720331 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 04:56:23 crc kubenswrapper[4799]: I1129 04:56:23.911509 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de975367-f439-4788-8ddb-b9f8af482589-log-httpd\") pod \"de975367-f439-4788-8ddb-b9f8af482589\" (UID: \"de975367-f439-4788-8ddb-b9f8af482589\") " Nov 29 04:56:23 crc kubenswrapper[4799]: I1129 04:56:23.911571 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de975367-f439-4788-8ddb-b9f8af482589-run-httpd\") pod \"de975367-f439-4788-8ddb-b9f8af482589\" (UID: \"de975367-f439-4788-8ddb-b9f8af482589\") " Nov 29 04:56:23 crc kubenswrapper[4799]: I1129 04:56:23.911623 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f2d22\" (UniqueName: \"kubernetes.io/projected/de975367-f439-4788-8ddb-b9f8af482589-kube-api-access-f2d22\") pod \"de975367-f439-4788-8ddb-b9f8af482589\" (UID: \"de975367-f439-4788-8ddb-b9f8af482589\") " Nov 29 04:56:23 crc kubenswrapper[4799]: I1129 04:56:23.911653 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/de975367-f439-4788-8ddb-b9f8af482589-sg-core-conf-yaml\") pod \"de975367-f439-4788-8ddb-b9f8af482589\" (UID: \"de975367-f439-4788-8ddb-b9f8af482589\") " Nov 29 04:56:23 crc kubenswrapper[4799]: I1129 04:56:23.911676 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de975367-f439-4788-8ddb-b9f8af482589-combined-ca-bundle\") pod \"de975367-f439-4788-8ddb-b9f8af482589\" (UID: \"de975367-f439-4788-8ddb-b9f8af482589\") " Nov 29 04:56:23 crc kubenswrapper[4799]: I1129 04:56:23.911724 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de975367-f439-4788-8ddb-b9f8af482589-scripts\") pod \"de975367-f439-4788-8ddb-b9f8af482589\" (UID: \"de975367-f439-4788-8ddb-b9f8af482589\") " Nov 29 04:56:23 crc kubenswrapper[4799]: I1129 04:56:23.911863 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de975367-f439-4788-8ddb-b9f8af482589-config-data\") pod \"de975367-f439-4788-8ddb-b9f8af482589\" (UID: \"de975367-f439-4788-8ddb-b9f8af482589\") " Nov 29 04:56:23 crc kubenswrapper[4799]: I1129 04:56:23.912329 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de975367-f439-4788-8ddb-b9f8af482589-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "de975367-f439-4788-8ddb-b9f8af482589" (UID: "de975367-f439-4788-8ddb-b9f8af482589"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:56:23 crc kubenswrapper[4799]: I1129 04:56:23.912425 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de975367-f439-4788-8ddb-b9f8af482589-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "de975367-f439-4788-8ddb-b9f8af482589" (UID: "de975367-f439-4788-8ddb-b9f8af482589"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:56:23 crc kubenswrapper[4799]: I1129 04:56:23.912840 4799 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de975367-f439-4788-8ddb-b9f8af482589-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:23 crc kubenswrapper[4799]: I1129 04:56:23.912862 4799 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de975367-f439-4788-8ddb-b9f8af482589-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:23 crc kubenswrapper[4799]: I1129 04:56:23.926196 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de975367-f439-4788-8ddb-b9f8af482589-scripts" (OuterVolumeSpecName: "scripts") pod "de975367-f439-4788-8ddb-b9f8af482589" (UID: "de975367-f439-4788-8ddb-b9f8af482589"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:56:23 crc kubenswrapper[4799]: I1129 04:56:23.929384 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de975367-f439-4788-8ddb-b9f8af482589-kube-api-access-f2d22" (OuterVolumeSpecName: "kube-api-access-f2d22") pod "de975367-f439-4788-8ddb-b9f8af482589" (UID: "de975367-f439-4788-8ddb-b9f8af482589"). InnerVolumeSpecName "kube-api-access-f2d22". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:56:23 crc kubenswrapper[4799]: I1129 04:56:23.954931 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de975367-f439-4788-8ddb-b9f8af482589-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "de975367-f439-4788-8ddb-b9f8af482589" (UID: "de975367-f439-4788-8ddb-b9f8af482589"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.008945 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de975367-f439-4788-8ddb-b9f8af482589-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "de975367-f439-4788-8ddb-b9f8af482589" (UID: "de975367-f439-4788-8ddb-b9f8af482589"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.014744 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de975367-f439-4788-8ddb-b9f8af482589-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.014807 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f2d22\" (UniqueName: \"kubernetes.io/projected/de975367-f439-4788-8ddb-b9f8af482589-kube-api-access-f2d22\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.014827 4799 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/de975367-f439-4788-8ddb-b9f8af482589-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.014842 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de975367-f439-4788-8ddb-b9f8af482589-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.047181 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-wq84l"] Nov 29 04:56:24 crc kubenswrapper[4799]: E1129 04:56:24.047760 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de975367-f439-4788-8ddb-b9f8af482589" containerName="sg-core" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.047810 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="de975367-f439-4788-8ddb-b9f8af482589" containerName="sg-core" Nov 29 04:56:24 crc kubenswrapper[4799]: E1129 04:56:24.047840 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de975367-f439-4788-8ddb-b9f8af482589" containerName="ceilometer-central-agent" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.047851 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="de975367-f439-4788-8ddb-b9f8af482589" containerName="ceilometer-central-agent" Nov 29 04:56:24 crc kubenswrapper[4799]: E1129 04:56:24.047871 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de975367-f439-4788-8ddb-b9f8af482589" containerName="proxy-httpd" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.047881 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="de975367-f439-4788-8ddb-b9f8af482589" containerName="proxy-httpd" Nov 29 04:56:24 crc kubenswrapper[4799]: E1129 04:56:24.047914 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de975367-f439-4788-8ddb-b9f8af482589" containerName="ceilometer-notification-agent" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.047922 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="de975367-f439-4788-8ddb-b9f8af482589" containerName="ceilometer-notification-agent" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.048152 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="de975367-f439-4788-8ddb-b9f8af482589" containerName="sg-core" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.048189 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="de975367-f439-4788-8ddb-b9f8af482589" containerName="ceilometer-notification-agent" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.048206 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="de975367-f439-4788-8ddb-b9f8af482589" containerName="ceilometer-central-agent" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.048220 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="de975367-f439-4788-8ddb-b9f8af482589" containerName="proxy-httpd" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.049126 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-wq84l" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.062172 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-wq84l"] Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.068808 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de975367-f439-4788-8ddb-b9f8af482589-config-data" (OuterVolumeSpecName: "config-data") pod "de975367-f439-4788-8ddb-b9f8af482589" (UID: "de975367-f439-4788-8ddb-b9f8af482589"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.116385 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5b670080-8ca4-4e80-a1a6-ec7a63c951b6-operator-scripts\") pod \"nova-api-db-create-wq84l\" (UID: \"5b670080-8ca4-4e80-a1a6-ec7a63c951b6\") " pod="openstack/nova-api-db-create-wq84l" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.116538 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c74ct\" (UniqueName: \"kubernetes.io/projected/5b670080-8ca4-4e80-a1a6-ec7a63c951b6-kube-api-access-c74ct\") pod \"nova-api-db-create-wq84l\" (UID: \"5b670080-8ca4-4e80-a1a6-ec7a63c951b6\") " pod="openstack/nova-api-db-create-wq84l" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.116664 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de975367-f439-4788-8ddb-b9f8af482589-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.138919 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-zcnn2"] Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.140342 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-zcnn2" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.147109 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-6872-account-create-update-ndcpk"] Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.148843 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-6872-account-create-update-ndcpk" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.152026 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.159704 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-zcnn2"] Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.175514 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-6872-account-create-update-ndcpk"] Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.218140 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5b670080-8ca4-4e80-a1a6-ec7a63c951b6-operator-scripts\") pod \"nova-api-db-create-wq84l\" (UID: \"5b670080-8ca4-4e80-a1a6-ec7a63c951b6\") " pod="openstack/nova-api-db-create-wq84l" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.218243 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c74ct\" (UniqueName: \"kubernetes.io/projected/5b670080-8ca4-4e80-a1a6-ec7a63c951b6-kube-api-access-c74ct\") pod \"nova-api-db-create-wq84l\" (UID: \"5b670080-8ca4-4e80-a1a6-ec7a63c951b6\") " pod="openstack/nova-api-db-create-wq84l" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.219229 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5b670080-8ca4-4e80-a1a6-ec7a63c951b6-operator-scripts\") pod \"nova-api-db-create-wq84l\" (UID: \"5b670080-8ca4-4e80-a1a6-ec7a63c951b6\") " pod="openstack/nova-api-db-create-wq84l" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.239576 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c74ct\" (UniqueName: \"kubernetes.io/projected/5b670080-8ca4-4e80-a1a6-ec7a63c951b6-kube-api-access-c74ct\") pod \"nova-api-db-create-wq84l\" (UID: \"5b670080-8ca4-4e80-a1a6-ec7a63c951b6\") " pod="openstack/nova-api-db-create-wq84l" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.319553 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/378ff2b6-e6b0-4cb0-a6ef-33e9ce1a7021-operator-scripts\") pod \"nova-cell0-db-create-zcnn2\" (UID: \"378ff2b6-e6b0-4cb0-a6ef-33e9ce1a7021\") " pod="openstack/nova-cell0-db-create-zcnn2" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.319802 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxz47\" (UniqueName: \"kubernetes.io/projected/d3bd96cd-ca29-4a82-8c97-0709b76cd021-kube-api-access-fxz47\") pod \"nova-api-6872-account-create-update-ndcpk\" (UID: \"d3bd96cd-ca29-4a82-8c97-0709b76cd021\") " pod="openstack/nova-api-6872-account-create-update-ndcpk" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.319863 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dllnv\" (UniqueName: \"kubernetes.io/projected/378ff2b6-e6b0-4cb0-a6ef-33e9ce1a7021-kube-api-access-dllnv\") pod \"nova-cell0-db-create-zcnn2\" (UID: \"378ff2b6-e6b0-4cb0-a6ef-33e9ce1a7021\") " pod="openstack/nova-cell0-db-create-zcnn2" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.319894 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3bd96cd-ca29-4a82-8c97-0709b76cd021-operator-scripts\") pod \"nova-api-6872-account-create-update-ndcpk\" (UID: \"d3bd96cd-ca29-4a82-8c97-0709b76cd021\") " pod="openstack/nova-api-6872-account-create-update-ndcpk" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.342392 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-hn8xp"] Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.343657 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-hn8xp" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.357968 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-48c8-account-create-update-t6p84"] Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.359433 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-48c8-account-create-update-t6p84" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.363328 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.369407 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-hn8xp"] Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.371437 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-wq84l" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.382451 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-48c8-account-create-update-t6p84"] Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.424143 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxz47\" (UniqueName: \"kubernetes.io/projected/d3bd96cd-ca29-4a82-8c97-0709b76cd021-kube-api-access-fxz47\") pod \"nova-api-6872-account-create-update-ndcpk\" (UID: \"d3bd96cd-ca29-4a82-8c97-0709b76cd021\") " pod="openstack/nova-api-6872-account-create-update-ndcpk" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.424229 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dllnv\" (UniqueName: \"kubernetes.io/projected/378ff2b6-e6b0-4cb0-a6ef-33e9ce1a7021-kube-api-access-dllnv\") pod \"nova-cell0-db-create-zcnn2\" (UID: \"378ff2b6-e6b0-4cb0-a6ef-33e9ce1a7021\") " pod="openstack/nova-cell0-db-create-zcnn2" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.424251 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3bd96cd-ca29-4a82-8c97-0709b76cd021-operator-scripts\") pod \"nova-api-6872-account-create-update-ndcpk\" (UID: \"d3bd96cd-ca29-4a82-8c97-0709b76cd021\") " pod="openstack/nova-api-6872-account-create-update-ndcpk" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.424299 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/378ff2b6-e6b0-4cb0-a6ef-33e9ce1a7021-operator-scripts\") pod \"nova-cell0-db-create-zcnn2\" (UID: \"378ff2b6-e6b0-4cb0-a6ef-33e9ce1a7021\") " pod="openstack/nova-cell0-db-create-zcnn2" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.425276 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/378ff2b6-e6b0-4cb0-a6ef-33e9ce1a7021-operator-scripts\") pod \"nova-cell0-db-create-zcnn2\" (UID: \"378ff2b6-e6b0-4cb0-a6ef-33e9ce1a7021\") " pod="openstack/nova-cell0-db-create-zcnn2" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.425917 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3bd96cd-ca29-4a82-8c97-0709b76cd021-operator-scripts\") pod \"nova-api-6872-account-create-update-ndcpk\" (UID: \"d3bd96cd-ca29-4a82-8c97-0709b76cd021\") " pod="openstack/nova-api-6872-account-create-update-ndcpk" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.454606 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxz47\" (UniqueName: \"kubernetes.io/projected/d3bd96cd-ca29-4a82-8c97-0709b76cd021-kube-api-access-fxz47\") pod \"nova-api-6872-account-create-update-ndcpk\" (UID: \"d3bd96cd-ca29-4a82-8c97-0709b76cd021\") " pod="openstack/nova-api-6872-account-create-update-ndcpk" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.460514 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dllnv\" (UniqueName: \"kubernetes.io/projected/378ff2b6-e6b0-4cb0-a6ef-33e9ce1a7021-kube-api-access-dllnv\") pod \"nova-cell0-db-create-zcnn2\" (UID: \"378ff2b6-e6b0-4cb0-a6ef-33e9ce1a7021\") " pod="openstack/nova-cell0-db-create-zcnn2" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.466408 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-zcnn2" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.483320 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-6872-account-create-update-ndcpk" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.527172 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjffp\" (UniqueName: \"kubernetes.io/projected/3452c4a6-ce88-4908-9ebf-6b53ccfcdc2e-kube-api-access-qjffp\") pod \"nova-cell0-48c8-account-create-update-t6p84\" (UID: \"3452c4a6-ce88-4908-9ebf-6b53ccfcdc2e\") " pod="openstack/nova-cell0-48c8-account-create-update-t6p84" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.527263 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3452c4a6-ce88-4908-9ebf-6b53ccfcdc2e-operator-scripts\") pod \"nova-cell0-48c8-account-create-update-t6p84\" (UID: \"3452c4a6-ce88-4908-9ebf-6b53ccfcdc2e\") " pod="openstack/nova-cell0-48c8-account-create-update-t6p84" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.527300 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbrrl\" (UniqueName: \"kubernetes.io/projected/cf5b318f-e3e1-45e0-b4d5-dd7fe656676f-kube-api-access-rbrrl\") pod \"nova-cell1-db-create-hn8xp\" (UID: \"cf5b318f-e3e1-45e0-b4d5-dd7fe656676f\") " pod="openstack/nova-cell1-db-create-hn8xp" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.527392 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf5b318f-e3e1-45e0-b4d5-dd7fe656676f-operator-scripts\") pod \"nova-cell1-db-create-hn8xp\" (UID: \"cf5b318f-e3e1-45e0-b4d5-dd7fe656676f\") " pod="openstack/nova-cell1-db-create-hn8xp" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.560336 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"548ff572-b2d0-45ec-a131-0ba87044b4ff","Type":"ContainerStarted","Data":"8c0caf9042172375d4cc2739bf1299fc5e481de4fe366642b60e3991aecbd4a3"} Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.578205 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de975367-f439-4788-8ddb-b9f8af482589","Type":"ContainerDied","Data":"3aa3cf1b4350391a08cc4f094520d74e14370c0def867d3fd71492f1dc600631"} Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.578479 4799 scope.go:117] "RemoveContainer" containerID="19f4901b0a5fd72073558e75cacf3cbb37d1bf2c2f64f0d5fb7ffc20577f1519" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.578488 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.579599 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-da5a-account-create-update-ntzr7"] Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.581095 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-da5a-account-create-update-ntzr7" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.595448 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-da5a-account-create-update-ntzr7"] Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.600070 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.602441 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.539207439 podStartE2EDuration="13.602418809s" podCreationTimestamp="2025-11-29 04:56:11 +0000 UTC" firstStartedPulling="2025-11-29 04:56:12.371366972 +0000 UTC m=+1048.014297372" lastFinishedPulling="2025-11-29 04:56:23.434578342 +0000 UTC m=+1059.077508742" observedRunningTime="2025-11-29 04:56:24.599391355 +0000 UTC m=+1060.242321755" watchObservedRunningTime="2025-11-29 04:56:24.602418809 +0000 UTC m=+1060.245349209" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.630375 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjffp\" (UniqueName: \"kubernetes.io/projected/3452c4a6-ce88-4908-9ebf-6b53ccfcdc2e-kube-api-access-qjffp\") pod \"nova-cell0-48c8-account-create-update-t6p84\" (UID: \"3452c4a6-ce88-4908-9ebf-6b53ccfcdc2e\") " pod="openstack/nova-cell0-48c8-account-create-update-t6p84" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.630450 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbrrl\" (UniqueName: \"kubernetes.io/projected/cf5b318f-e3e1-45e0-b4d5-dd7fe656676f-kube-api-access-rbrrl\") pod \"nova-cell1-db-create-hn8xp\" (UID: \"cf5b318f-e3e1-45e0-b4d5-dd7fe656676f\") " pod="openstack/nova-cell1-db-create-hn8xp" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.630472 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3452c4a6-ce88-4908-9ebf-6b53ccfcdc2e-operator-scripts\") pod \"nova-cell0-48c8-account-create-update-t6p84\" (UID: \"3452c4a6-ce88-4908-9ebf-6b53ccfcdc2e\") " pod="openstack/nova-cell0-48c8-account-create-update-t6p84" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.630550 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf5b318f-e3e1-45e0-b4d5-dd7fe656676f-operator-scripts\") pod \"nova-cell1-db-create-hn8xp\" (UID: \"cf5b318f-e3e1-45e0-b4d5-dd7fe656676f\") " pod="openstack/nova-cell1-db-create-hn8xp" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.631757 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf5b318f-e3e1-45e0-b4d5-dd7fe656676f-operator-scripts\") pod \"nova-cell1-db-create-hn8xp\" (UID: \"cf5b318f-e3e1-45e0-b4d5-dd7fe656676f\") " pod="openstack/nova-cell1-db-create-hn8xp" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.632185 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3452c4a6-ce88-4908-9ebf-6b53ccfcdc2e-operator-scripts\") pod \"nova-cell0-48c8-account-create-update-t6p84\" (UID: \"3452c4a6-ce88-4908-9ebf-6b53ccfcdc2e\") " pod="openstack/nova-cell0-48c8-account-create-update-t6p84" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.650618 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjffp\" (UniqueName: \"kubernetes.io/projected/3452c4a6-ce88-4908-9ebf-6b53ccfcdc2e-kube-api-access-qjffp\") pod \"nova-cell0-48c8-account-create-update-t6p84\" (UID: \"3452c4a6-ce88-4908-9ebf-6b53ccfcdc2e\") " pod="openstack/nova-cell0-48c8-account-create-update-t6p84" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.652760 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbrrl\" (UniqueName: \"kubernetes.io/projected/cf5b318f-e3e1-45e0-b4d5-dd7fe656676f-kube-api-access-rbrrl\") pod \"nova-cell1-db-create-hn8xp\" (UID: \"cf5b318f-e3e1-45e0-b4d5-dd7fe656676f\") " pod="openstack/nova-cell1-db-create-hn8xp" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.700675 4799 scope.go:117] "RemoveContainer" containerID="22eed2795fed04c2ad4738a5b7ddd275701000eb0b9c938ced931a138b12ab81" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.707614 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-hn8xp" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.727313 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-48c8-account-create-update-t6p84" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.743449 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89glt\" (UniqueName: \"kubernetes.io/projected/b88ce1ac-1c8d-407d-8183-cdac9f354607-kube-api-access-89glt\") pod \"nova-cell1-da5a-account-create-update-ntzr7\" (UID: \"b88ce1ac-1c8d-407d-8183-cdac9f354607\") " pod="openstack/nova-cell1-da5a-account-create-update-ntzr7" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.745077 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b88ce1ac-1c8d-407d-8183-cdac9f354607-operator-scripts\") pod \"nova-cell1-da5a-account-create-update-ntzr7\" (UID: \"b88ce1ac-1c8d-407d-8183-cdac9f354607\") " pod="openstack/nova-cell1-da5a-account-create-update-ntzr7" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.779026 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.798076 4799 scope.go:117] "RemoveContainer" containerID="fbd3059919aa4e1d1d2943fd052b2861ed1b731729d22d5383486cc878702a33" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.811011 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.818414 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.820626 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.831374 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.831815 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.846593 4799 scope.go:117] "RemoveContainer" containerID="59ea15e9044a8c7f1247d7fb32c1ad6ab35c795668363a83216a4a7e201564c0" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.851009 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b88ce1ac-1c8d-407d-8183-cdac9f354607-operator-scripts\") pod \"nova-cell1-da5a-account-create-update-ntzr7\" (UID: \"b88ce1ac-1c8d-407d-8183-cdac9f354607\") " pod="openstack/nova-cell1-da5a-account-create-update-ntzr7" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.851234 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89glt\" (UniqueName: \"kubernetes.io/projected/b88ce1ac-1c8d-407d-8183-cdac9f354607-kube-api-access-89glt\") pod \"nova-cell1-da5a-account-create-update-ntzr7\" (UID: \"b88ce1ac-1c8d-407d-8183-cdac9f354607\") " pod="openstack/nova-cell1-da5a-account-create-update-ntzr7" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.853733 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b88ce1ac-1c8d-407d-8183-cdac9f354607-operator-scripts\") pod \"nova-cell1-da5a-account-create-update-ntzr7\" (UID: \"b88ce1ac-1c8d-407d-8183-cdac9f354607\") " pod="openstack/nova-cell1-da5a-account-create-update-ntzr7" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.869863 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.870446 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89glt\" (UniqueName: \"kubernetes.io/projected/b88ce1ac-1c8d-407d-8183-cdac9f354607-kube-api-access-89glt\") pod \"nova-cell1-da5a-account-create-update-ntzr7\" (UID: \"b88ce1ac-1c8d-407d-8183-cdac9f354607\") " pod="openstack/nova-cell1-da5a-account-create-update-ntzr7" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.952850 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fn2td\" (UniqueName: \"kubernetes.io/projected/e97e3977-fa5b-4723-9dcc-55907243890d-kube-api-access-fn2td\") pod \"ceilometer-0\" (UID: \"e97e3977-fa5b-4723-9dcc-55907243890d\") " pod="openstack/ceilometer-0" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.952930 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e97e3977-fa5b-4723-9dcc-55907243890d-config-data\") pod \"ceilometer-0\" (UID: \"e97e3977-fa5b-4723-9dcc-55907243890d\") " pod="openstack/ceilometer-0" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.952994 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e97e3977-fa5b-4723-9dcc-55907243890d-log-httpd\") pod \"ceilometer-0\" (UID: \"e97e3977-fa5b-4723-9dcc-55907243890d\") " pod="openstack/ceilometer-0" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.953016 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e97e3977-fa5b-4723-9dcc-55907243890d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e97e3977-fa5b-4723-9dcc-55907243890d\") " pod="openstack/ceilometer-0" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.953046 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e97e3977-fa5b-4723-9dcc-55907243890d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e97e3977-fa5b-4723-9dcc-55907243890d\") " pod="openstack/ceilometer-0" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.953107 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e97e3977-fa5b-4723-9dcc-55907243890d-scripts\") pod \"ceilometer-0\" (UID: \"e97e3977-fa5b-4723-9dcc-55907243890d\") " pod="openstack/ceilometer-0" Nov 29 04:56:24 crc kubenswrapper[4799]: I1129 04:56:24.953214 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e97e3977-fa5b-4723-9dcc-55907243890d-run-httpd\") pod \"ceilometer-0\" (UID: \"e97e3977-fa5b-4723-9dcc-55907243890d\") " pod="openstack/ceilometer-0" Nov 29 04:56:25 crc kubenswrapper[4799]: I1129 04:56:25.013397 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-wq84l"] Nov 29 04:56:25 crc kubenswrapper[4799]: W1129 04:56:25.021594 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b670080_8ca4_4e80_a1a6_ec7a63c951b6.slice/crio-60a83bc58f9bddb9bf5d9e9acbe836faadef07dabfc04886d2b6bd78d589ad25 WatchSource:0}: Error finding container 60a83bc58f9bddb9bf5d9e9acbe836faadef07dabfc04886d2b6bd78d589ad25: Status 404 returned error can't find the container with id 60a83bc58f9bddb9bf5d9e9acbe836faadef07dabfc04886d2b6bd78d589ad25 Nov 29 04:56:25 crc kubenswrapper[4799]: I1129 04:56:25.035010 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-da5a-account-create-update-ntzr7" Nov 29 04:56:25 crc kubenswrapper[4799]: I1129 04:56:25.054816 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e97e3977-fa5b-4723-9dcc-55907243890d-run-httpd\") pod \"ceilometer-0\" (UID: \"e97e3977-fa5b-4723-9dcc-55907243890d\") " pod="openstack/ceilometer-0" Nov 29 04:56:25 crc kubenswrapper[4799]: I1129 04:56:25.054889 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fn2td\" (UniqueName: \"kubernetes.io/projected/e97e3977-fa5b-4723-9dcc-55907243890d-kube-api-access-fn2td\") pod \"ceilometer-0\" (UID: \"e97e3977-fa5b-4723-9dcc-55907243890d\") " pod="openstack/ceilometer-0" Nov 29 04:56:25 crc kubenswrapper[4799]: I1129 04:56:25.054929 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e97e3977-fa5b-4723-9dcc-55907243890d-config-data\") pod \"ceilometer-0\" (UID: \"e97e3977-fa5b-4723-9dcc-55907243890d\") " pod="openstack/ceilometer-0" Nov 29 04:56:25 crc kubenswrapper[4799]: I1129 04:56:25.054975 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e97e3977-fa5b-4723-9dcc-55907243890d-log-httpd\") pod \"ceilometer-0\" (UID: \"e97e3977-fa5b-4723-9dcc-55907243890d\") " pod="openstack/ceilometer-0" Nov 29 04:56:25 crc kubenswrapper[4799]: I1129 04:56:25.054995 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e97e3977-fa5b-4723-9dcc-55907243890d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e97e3977-fa5b-4723-9dcc-55907243890d\") " pod="openstack/ceilometer-0" Nov 29 04:56:25 crc kubenswrapper[4799]: I1129 04:56:25.055019 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e97e3977-fa5b-4723-9dcc-55907243890d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e97e3977-fa5b-4723-9dcc-55907243890d\") " pod="openstack/ceilometer-0" Nov 29 04:56:25 crc kubenswrapper[4799]: I1129 04:56:25.055072 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e97e3977-fa5b-4723-9dcc-55907243890d-scripts\") pod \"ceilometer-0\" (UID: \"e97e3977-fa5b-4723-9dcc-55907243890d\") " pod="openstack/ceilometer-0" Nov 29 04:56:25 crc kubenswrapper[4799]: I1129 04:56:25.055650 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e97e3977-fa5b-4723-9dcc-55907243890d-run-httpd\") pod \"ceilometer-0\" (UID: \"e97e3977-fa5b-4723-9dcc-55907243890d\") " pod="openstack/ceilometer-0" Nov 29 04:56:25 crc kubenswrapper[4799]: I1129 04:56:25.055933 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e97e3977-fa5b-4723-9dcc-55907243890d-log-httpd\") pod \"ceilometer-0\" (UID: \"e97e3977-fa5b-4723-9dcc-55907243890d\") " pod="openstack/ceilometer-0" Nov 29 04:56:25 crc kubenswrapper[4799]: I1129 04:56:25.062684 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e97e3977-fa5b-4723-9dcc-55907243890d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e97e3977-fa5b-4723-9dcc-55907243890d\") " pod="openstack/ceilometer-0" Nov 29 04:56:25 crc kubenswrapper[4799]: I1129 04:56:25.063449 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e97e3977-fa5b-4723-9dcc-55907243890d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e97e3977-fa5b-4723-9dcc-55907243890d\") " pod="openstack/ceilometer-0" Nov 29 04:56:25 crc kubenswrapper[4799]: I1129 04:56:25.066236 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e97e3977-fa5b-4723-9dcc-55907243890d-config-data\") pod \"ceilometer-0\" (UID: \"e97e3977-fa5b-4723-9dcc-55907243890d\") " pod="openstack/ceilometer-0" Nov 29 04:56:25 crc kubenswrapper[4799]: I1129 04:56:25.072108 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e97e3977-fa5b-4723-9dcc-55907243890d-scripts\") pod \"ceilometer-0\" (UID: \"e97e3977-fa5b-4723-9dcc-55907243890d\") " pod="openstack/ceilometer-0" Nov 29 04:56:25 crc kubenswrapper[4799]: I1129 04:56:25.073622 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fn2td\" (UniqueName: \"kubernetes.io/projected/e97e3977-fa5b-4723-9dcc-55907243890d-kube-api-access-fn2td\") pod \"ceilometer-0\" (UID: \"e97e3977-fa5b-4723-9dcc-55907243890d\") " pod="openstack/ceilometer-0" Nov 29 04:56:25 crc kubenswrapper[4799]: I1129 04:56:25.144396 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-zcnn2"] Nov 29 04:56:25 crc kubenswrapper[4799]: I1129 04:56:25.166864 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 04:56:25 crc kubenswrapper[4799]: I1129 04:56:25.283938 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-6872-account-create-update-ndcpk"] Nov 29 04:56:25 crc kubenswrapper[4799]: I1129 04:56:25.398111 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-48c8-account-create-update-t6p84"] Nov 29 04:56:25 crc kubenswrapper[4799]: I1129 04:56:25.492124 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-hn8xp"] Nov 29 04:56:25 crc kubenswrapper[4799]: I1129 04:56:25.587207 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-wq84l" event={"ID":"5b670080-8ca4-4e80-a1a6-ec7a63c951b6","Type":"ContainerStarted","Data":"60a83bc58f9bddb9bf5d9e9acbe836faadef07dabfc04886d2b6bd78d589ad25"} Nov 29 04:56:25 crc kubenswrapper[4799]: I1129 04:56:25.589551 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-zcnn2" event={"ID":"378ff2b6-e6b0-4cb0-a6ef-33e9ce1a7021","Type":"ContainerStarted","Data":"9bf1a61ddf57c2567f3638d3f5eab042c6c76a057190d6721b3824551befc6c0"} Nov 29 04:56:25 crc kubenswrapper[4799]: I1129 04:56:25.590850 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-hn8xp" event={"ID":"cf5b318f-e3e1-45e0-b4d5-dd7fe656676f","Type":"ContainerStarted","Data":"45854e1296321401f23fca00a556958adcaf2c47d25ba35f94c0614e0ecde256"} Nov 29 04:56:25 crc kubenswrapper[4799]: I1129 04:56:25.591829 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-48c8-account-create-update-t6p84" event={"ID":"3452c4a6-ce88-4908-9ebf-6b53ccfcdc2e","Type":"ContainerStarted","Data":"da5fd9e5487f0a4446471475dee93e2a83d70dbacdaac9e9a4231d2285bb023c"} Nov 29 04:56:25 crc kubenswrapper[4799]: I1129 04:56:25.594576 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-6872-account-create-update-ndcpk" event={"ID":"d3bd96cd-ca29-4a82-8c97-0709b76cd021","Type":"ContainerStarted","Data":"ab3103cfd6dde264bd6985d98d14271eb93b39b91b6a037feeed1a9c009d2589"} Nov 29 04:56:25 crc kubenswrapper[4799]: I1129 04:56:25.639577 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 29 04:56:25 crc kubenswrapper[4799]: I1129 04:56:25.640873 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-da5a-account-create-update-ntzr7"] Nov 29 04:56:25 crc kubenswrapper[4799]: I1129 04:56:25.729656 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 29 04:56:25 crc kubenswrapper[4799]: W1129 04:56:25.751988 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode97e3977_fa5b_4723_9dcc_55907243890d.slice/crio-2b99e42389aa495845f6d8c05f71ea3208c64ae83ba7bb4fc0f4909d39a4e449 WatchSource:0}: Error finding container 2b99e42389aa495845f6d8c05f71ea3208c64ae83ba7bb4fc0f4909d39a4e449: Status 404 returned error can't find the container with id 2b99e42389aa495845f6d8c05f71ea3208c64ae83ba7bb4fc0f4909d39a4e449 Nov 29 04:56:26 crc kubenswrapper[4799]: I1129 04:56:26.642194 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e97e3977-fa5b-4723-9dcc-55907243890d","Type":"ContainerStarted","Data":"2b99e42389aa495845f6d8c05f71ea3208c64ae83ba7bb4fc0f4909d39a4e449"} Nov 29 04:56:26 crc kubenswrapper[4799]: I1129 04:56:26.647930 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-da5a-account-create-update-ntzr7" event={"ID":"b88ce1ac-1c8d-407d-8183-cdac9f354607","Type":"ContainerStarted","Data":"aa4dd2af37bb6eb1b45670f388f98036191d2ca766e280c90fdfd4fdf121f0cf"} Nov 29 04:56:26 crc kubenswrapper[4799]: I1129 04:56:26.669545 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de975367-f439-4788-8ddb-b9f8af482589" path="/var/lib/kubelet/pods/de975367-f439-4788-8ddb-b9f8af482589/volumes" Nov 29 04:56:27 crc kubenswrapper[4799]: I1129 04:56:27.658243 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-48c8-account-create-update-t6p84" event={"ID":"3452c4a6-ce88-4908-9ebf-6b53ccfcdc2e","Type":"ContainerStarted","Data":"55c318edd4521e50535e4e9250780b650068d4c774473c1d6f4882bdfa1245ae"} Nov 29 04:56:27 crc kubenswrapper[4799]: I1129 04:56:27.662393 4799 generic.go:334] "Generic (PLEG): container finished" podID="f21cec8f-8cd4-4ecd-9800-3bb0c482fb66" containerID="0206e122fb25f53bced3ad79d205461cb91bdecdbdf24990b6664f0037824b4c" exitCode=0 Nov 29 04:56:27 crc kubenswrapper[4799]: I1129 04:56:27.662414 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d557d86d8-6zk7x" event={"ID":"f21cec8f-8cd4-4ecd-9800-3bb0c482fb66","Type":"ContainerDied","Data":"0206e122fb25f53bced3ad79d205461cb91bdecdbdf24990b6664f0037824b4c"} Nov 29 04:56:27 crc kubenswrapper[4799]: I1129 04:56:27.664237 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e97e3977-fa5b-4723-9dcc-55907243890d","Type":"ContainerStarted","Data":"f7817ff369dcbebba92bf72b341244eae9027d4b2c5da4c4c22a39f691bb3040"} Nov 29 04:56:27 crc kubenswrapper[4799]: I1129 04:56:27.665962 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-6872-account-create-update-ndcpk" event={"ID":"d3bd96cd-ca29-4a82-8c97-0709b76cd021","Type":"ContainerStarted","Data":"2dd4967ea72c729705b7523ff2b93e3133362ad86b2a98be72716ef1eb1b4ab6"} Nov 29 04:56:27 crc kubenswrapper[4799]: I1129 04:56:27.667404 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-wq84l" event={"ID":"5b670080-8ca4-4e80-a1a6-ec7a63c951b6","Type":"ContainerStarted","Data":"9c14caf55dc6373fc47f3c065a099730372a3a88825444471218dabaa64b18d9"} Nov 29 04:56:27 crc kubenswrapper[4799]: I1129 04:56:27.668752 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-zcnn2" event={"ID":"378ff2b6-e6b0-4cb0-a6ef-33e9ce1a7021","Type":"ContainerStarted","Data":"87cab5ee995c7c6d7b9159edcebe0c1a700965ef558a9dbd8ac8a6e1ab8b8730"} Nov 29 04:56:27 crc kubenswrapper[4799]: I1129 04:56:27.669968 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-hn8xp" event={"ID":"cf5b318f-e3e1-45e0-b4d5-dd7fe656676f","Type":"ContainerStarted","Data":"8674cdb7fd19ef3fdbb881b1c8e30eb93da76ed230a3598ef0d109c29ebffcf8"} Nov 29 04:56:27 crc kubenswrapper[4799]: I1129 04:56:27.675114 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-48c8-account-create-update-t6p84" podStartSLOduration=3.675099617 podStartE2EDuration="3.675099617s" podCreationTimestamp="2025-11-29 04:56:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:56:27.67356472 +0000 UTC m=+1063.316495120" watchObservedRunningTime="2025-11-29 04:56:27.675099617 +0000 UTC m=+1063.318030017" Nov 29 04:56:27 crc kubenswrapper[4799]: I1129 04:56:27.680285 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-da5a-account-create-update-ntzr7" event={"ID":"b88ce1ac-1c8d-407d-8183-cdac9f354607","Type":"ContainerStarted","Data":"d563bee704f704153e345a733a1a537d1c957e8507f87cf5adef9476f58065a8"} Nov 29 04:56:27 crc kubenswrapper[4799]: I1129 04:56:27.697201 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-6872-account-create-update-ndcpk" podStartSLOduration=3.6971784149999998 podStartE2EDuration="3.697178415s" podCreationTimestamp="2025-11-29 04:56:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:56:27.691136148 +0000 UTC m=+1063.334066558" watchObservedRunningTime="2025-11-29 04:56:27.697178415 +0000 UTC m=+1063.340108815" Nov 29 04:56:27 crc kubenswrapper[4799]: I1129 04:56:27.711571 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-hn8xp" podStartSLOduration=3.711541407 podStartE2EDuration="3.711541407s" podCreationTimestamp="2025-11-29 04:56:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:56:27.704994657 +0000 UTC m=+1063.347925057" watchObservedRunningTime="2025-11-29 04:56:27.711541407 +0000 UTC m=+1063.354471807" Nov 29 04:56:27 crc kubenswrapper[4799]: I1129 04:56:27.731583 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-zcnn2" podStartSLOduration=3.731560185 podStartE2EDuration="3.731560185s" podCreationTimestamp="2025-11-29 04:56:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:56:27.720211278 +0000 UTC m=+1063.363141698" watchObservedRunningTime="2025-11-29 04:56:27.731560185 +0000 UTC m=+1063.374490585" Nov 29 04:56:27 crc kubenswrapper[4799]: I1129 04:56:27.749260 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-wq84l" podStartSLOduration=3.749234426 podStartE2EDuration="3.749234426s" podCreationTimestamp="2025-11-29 04:56:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:56:27.737585982 +0000 UTC m=+1063.380516382" watchObservedRunningTime="2025-11-29 04:56:27.749234426 +0000 UTC m=+1063.392164826" Nov 29 04:56:27 crc kubenswrapper[4799]: I1129 04:56:27.756160 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-da5a-account-create-update-ntzr7" podStartSLOduration=3.756138254 podStartE2EDuration="3.756138254s" podCreationTimestamp="2025-11-29 04:56:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:56:27.752123777 +0000 UTC m=+1063.395054177" watchObservedRunningTime="2025-11-29 04:56:27.756138254 +0000 UTC m=+1063.399068654" Nov 29 04:56:28 crc kubenswrapper[4799]: I1129 04:56:28.691314 4799 generic.go:334] "Generic (PLEG): container finished" podID="3452c4a6-ce88-4908-9ebf-6b53ccfcdc2e" containerID="55c318edd4521e50535e4e9250780b650068d4c774473c1d6f4882bdfa1245ae" exitCode=0 Nov 29 04:56:28 crc kubenswrapper[4799]: I1129 04:56:28.691829 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-48c8-account-create-update-t6p84" event={"ID":"3452c4a6-ce88-4908-9ebf-6b53ccfcdc2e","Type":"ContainerDied","Data":"55c318edd4521e50535e4e9250780b650068d4c774473c1d6f4882bdfa1245ae"} Nov 29 04:56:28 crc kubenswrapper[4799]: I1129 04:56:28.699864 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d557d86d8-6zk7x" event={"ID":"f21cec8f-8cd4-4ecd-9800-3bb0c482fb66","Type":"ContainerDied","Data":"6ff3e3929129c98355795f8e80b1e4d66212c0f3d50ff9159807270bd2648c2d"} Nov 29 04:56:28 crc kubenswrapper[4799]: I1129 04:56:28.699911 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6ff3e3929129c98355795f8e80b1e4d66212c0f3d50ff9159807270bd2648c2d" Nov 29 04:56:28 crc kubenswrapper[4799]: I1129 04:56:28.701650 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e97e3977-fa5b-4723-9dcc-55907243890d","Type":"ContainerStarted","Data":"517be1f3d946d604be5a9dd7f7b645f08b271f832012ff40acbbab0df72a7702"} Nov 29 04:56:28 crc kubenswrapper[4799]: I1129 04:56:28.703205 4799 generic.go:334] "Generic (PLEG): container finished" podID="d3bd96cd-ca29-4a82-8c97-0709b76cd021" containerID="2dd4967ea72c729705b7523ff2b93e3133362ad86b2a98be72716ef1eb1b4ab6" exitCode=0 Nov 29 04:56:28 crc kubenswrapper[4799]: I1129 04:56:28.703261 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-6872-account-create-update-ndcpk" event={"ID":"d3bd96cd-ca29-4a82-8c97-0709b76cd021","Type":"ContainerDied","Data":"2dd4967ea72c729705b7523ff2b93e3133362ad86b2a98be72716ef1eb1b4ab6"} Nov 29 04:56:28 crc kubenswrapper[4799]: I1129 04:56:28.704598 4799 generic.go:334] "Generic (PLEG): container finished" podID="5b670080-8ca4-4e80-a1a6-ec7a63c951b6" containerID="9c14caf55dc6373fc47f3c065a099730372a3a88825444471218dabaa64b18d9" exitCode=0 Nov 29 04:56:28 crc kubenswrapper[4799]: I1129 04:56:28.704692 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-wq84l" event={"ID":"5b670080-8ca4-4e80-a1a6-ec7a63c951b6","Type":"ContainerDied","Data":"9c14caf55dc6373fc47f3c065a099730372a3a88825444471218dabaa64b18d9"} Nov 29 04:56:28 crc kubenswrapper[4799]: I1129 04:56:28.705943 4799 generic.go:334] "Generic (PLEG): container finished" podID="378ff2b6-e6b0-4cb0-a6ef-33e9ce1a7021" containerID="87cab5ee995c7c6d7b9159edcebe0c1a700965ef558a9dbd8ac8a6e1ab8b8730" exitCode=0 Nov 29 04:56:28 crc kubenswrapper[4799]: I1129 04:56:28.706012 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-zcnn2" event={"ID":"378ff2b6-e6b0-4cb0-a6ef-33e9ce1a7021","Type":"ContainerDied","Data":"87cab5ee995c7c6d7b9159edcebe0c1a700965ef558a9dbd8ac8a6e1ab8b8730"} Nov 29 04:56:28 crc kubenswrapper[4799]: I1129 04:56:28.707069 4799 generic.go:334] "Generic (PLEG): container finished" podID="cf5b318f-e3e1-45e0-b4d5-dd7fe656676f" containerID="8674cdb7fd19ef3fdbb881b1c8e30eb93da76ed230a3598ef0d109c29ebffcf8" exitCode=0 Nov 29 04:56:28 crc kubenswrapper[4799]: I1129 04:56:28.707125 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-hn8xp" event={"ID":"cf5b318f-e3e1-45e0-b4d5-dd7fe656676f","Type":"ContainerDied","Data":"8674cdb7fd19ef3fdbb881b1c8e30eb93da76ed230a3598ef0d109c29ebffcf8"} Nov 29 04:56:28 crc kubenswrapper[4799]: I1129 04:56:28.708329 4799 generic.go:334] "Generic (PLEG): container finished" podID="b88ce1ac-1c8d-407d-8183-cdac9f354607" containerID="d563bee704f704153e345a733a1a537d1c957e8507f87cf5adef9476f58065a8" exitCode=0 Nov 29 04:56:28 crc kubenswrapper[4799]: I1129 04:56:28.708395 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-da5a-account-create-update-ntzr7" event={"ID":"b88ce1ac-1c8d-407d-8183-cdac9f354607","Type":"ContainerDied","Data":"d563bee704f704153e345a733a1a537d1c957e8507f87cf5adef9476f58065a8"} Nov 29 04:56:28 crc kubenswrapper[4799]: I1129 04:56:28.758374 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d557d86d8-6zk7x" Nov 29 04:56:28 crc kubenswrapper[4799]: I1129 04:56:28.936715 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f21cec8f-8cd4-4ecd-9800-3bb0c482fb66-combined-ca-bundle\") pod \"f21cec8f-8cd4-4ecd-9800-3bb0c482fb66\" (UID: \"f21cec8f-8cd4-4ecd-9800-3bb0c482fb66\") " Nov 29 04:56:28 crc kubenswrapper[4799]: I1129 04:56:28.938313 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f21cec8f-8cd4-4ecd-9800-3bb0c482fb66-config\") pod \"f21cec8f-8cd4-4ecd-9800-3bb0c482fb66\" (UID: \"f21cec8f-8cd4-4ecd-9800-3bb0c482fb66\") " Nov 29 04:56:28 crc kubenswrapper[4799]: I1129 04:56:28.939008 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9hm44\" (UniqueName: \"kubernetes.io/projected/f21cec8f-8cd4-4ecd-9800-3bb0c482fb66-kube-api-access-9hm44\") pod \"f21cec8f-8cd4-4ecd-9800-3bb0c482fb66\" (UID: \"f21cec8f-8cd4-4ecd-9800-3bb0c482fb66\") " Nov 29 04:56:28 crc kubenswrapper[4799]: I1129 04:56:28.939106 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/f21cec8f-8cd4-4ecd-9800-3bb0c482fb66-httpd-config\") pod \"f21cec8f-8cd4-4ecd-9800-3bb0c482fb66\" (UID: \"f21cec8f-8cd4-4ecd-9800-3bb0c482fb66\") " Nov 29 04:56:28 crc kubenswrapper[4799]: I1129 04:56:28.939286 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f21cec8f-8cd4-4ecd-9800-3bb0c482fb66-ovndb-tls-certs\") pod \"f21cec8f-8cd4-4ecd-9800-3bb0c482fb66\" (UID: \"f21cec8f-8cd4-4ecd-9800-3bb0c482fb66\") " Nov 29 04:56:28 crc kubenswrapper[4799]: I1129 04:56:28.944410 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f21cec8f-8cd4-4ecd-9800-3bb0c482fb66-kube-api-access-9hm44" (OuterVolumeSpecName: "kube-api-access-9hm44") pod "f21cec8f-8cd4-4ecd-9800-3bb0c482fb66" (UID: "f21cec8f-8cd4-4ecd-9800-3bb0c482fb66"). InnerVolumeSpecName "kube-api-access-9hm44". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:56:28 crc kubenswrapper[4799]: I1129 04:56:28.944718 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f21cec8f-8cd4-4ecd-9800-3bb0c482fb66-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "f21cec8f-8cd4-4ecd-9800-3bb0c482fb66" (UID: "f21cec8f-8cd4-4ecd-9800-3bb0c482fb66"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:56:28 crc kubenswrapper[4799]: I1129 04:56:28.992330 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f21cec8f-8cd4-4ecd-9800-3bb0c482fb66-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f21cec8f-8cd4-4ecd-9800-3bb0c482fb66" (UID: "f21cec8f-8cd4-4ecd-9800-3bb0c482fb66"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:56:29 crc kubenswrapper[4799]: I1129 04:56:29.009123 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f21cec8f-8cd4-4ecd-9800-3bb0c482fb66-config" (OuterVolumeSpecName: "config") pod "f21cec8f-8cd4-4ecd-9800-3bb0c482fb66" (UID: "f21cec8f-8cd4-4ecd-9800-3bb0c482fb66"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:56:29 crc kubenswrapper[4799]: I1129 04:56:29.023812 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f21cec8f-8cd4-4ecd-9800-3bb0c482fb66-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "f21cec8f-8cd4-4ecd-9800-3bb0c482fb66" (UID: "f21cec8f-8cd4-4ecd-9800-3bb0c482fb66"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:56:29 crc kubenswrapper[4799]: I1129 04:56:29.042181 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/f21cec8f-8cd4-4ecd-9800-3bb0c482fb66-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:29 crc kubenswrapper[4799]: I1129 04:56:29.042222 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9hm44\" (UniqueName: \"kubernetes.io/projected/f21cec8f-8cd4-4ecd-9800-3bb0c482fb66-kube-api-access-9hm44\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:29 crc kubenswrapper[4799]: I1129 04:56:29.042240 4799 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/f21cec8f-8cd4-4ecd-9800-3bb0c482fb66-httpd-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:29 crc kubenswrapper[4799]: I1129 04:56:29.042253 4799 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f21cec8f-8cd4-4ecd-9800-3bb0c482fb66-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:29 crc kubenswrapper[4799]: I1129 04:56:29.042268 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f21cec8f-8cd4-4ecd-9800-3bb0c482fb66-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:29 crc kubenswrapper[4799]: I1129 04:56:29.405693 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 29 04:56:29 crc kubenswrapper[4799]: I1129 04:56:29.719320 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e97e3977-fa5b-4723-9dcc-55907243890d","Type":"ContainerStarted","Data":"7ad491dbe6e90a0e0aeeb6bcba1bbe8615e3a298443179011c8544976f56b561"} Nov 29 04:56:29 crc kubenswrapper[4799]: I1129 04:56:29.719727 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d557d86d8-6zk7x" Nov 29 04:56:29 crc kubenswrapper[4799]: I1129 04:56:29.799780 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-d557d86d8-6zk7x"] Nov 29 04:56:29 crc kubenswrapper[4799]: I1129 04:56:29.825560 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-d557d86d8-6zk7x"] Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.203703 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-48c8-account-create-update-t6p84" Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.379219 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3452c4a6-ce88-4908-9ebf-6b53ccfcdc2e-operator-scripts\") pod \"3452c4a6-ce88-4908-9ebf-6b53ccfcdc2e\" (UID: \"3452c4a6-ce88-4908-9ebf-6b53ccfcdc2e\") " Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.379295 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qjffp\" (UniqueName: \"kubernetes.io/projected/3452c4a6-ce88-4908-9ebf-6b53ccfcdc2e-kube-api-access-qjffp\") pod \"3452c4a6-ce88-4908-9ebf-6b53ccfcdc2e\" (UID: \"3452c4a6-ce88-4908-9ebf-6b53ccfcdc2e\") " Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.379875 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3452c4a6-ce88-4908-9ebf-6b53ccfcdc2e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3452c4a6-ce88-4908-9ebf-6b53ccfcdc2e" (UID: "3452c4a6-ce88-4908-9ebf-6b53ccfcdc2e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.397728 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3452c4a6-ce88-4908-9ebf-6b53ccfcdc2e-kube-api-access-qjffp" (OuterVolumeSpecName: "kube-api-access-qjffp") pod "3452c4a6-ce88-4908-9ebf-6b53ccfcdc2e" (UID: "3452c4a6-ce88-4908-9ebf-6b53ccfcdc2e"). InnerVolumeSpecName "kube-api-access-qjffp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.447851 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-wq84l" Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.454157 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-da5a-account-create-update-ntzr7" Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.459896 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-zcnn2" Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.465466 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-6872-account-create-update-ndcpk" Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.478290 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-hn8xp" Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.480568 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c74ct\" (UniqueName: \"kubernetes.io/projected/5b670080-8ca4-4e80-a1a6-ec7a63c951b6-kube-api-access-c74ct\") pod \"5b670080-8ca4-4e80-a1a6-ec7a63c951b6\" (UID: \"5b670080-8ca4-4e80-a1a6-ec7a63c951b6\") " Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.480630 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbrrl\" (UniqueName: \"kubernetes.io/projected/cf5b318f-e3e1-45e0-b4d5-dd7fe656676f-kube-api-access-rbrrl\") pod \"cf5b318f-e3e1-45e0-b4d5-dd7fe656676f\" (UID: \"cf5b318f-e3e1-45e0-b4d5-dd7fe656676f\") " Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.480663 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/378ff2b6-e6b0-4cb0-a6ef-33e9ce1a7021-operator-scripts\") pod \"378ff2b6-e6b0-4cb0-a6ef-33e9ce1a7021\" (UID: \"378ff2b6-e6b0-4cb0-a6ef-33e9ce1a7021\") " Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.480698 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-89glt\" (UniqueName: \"kubernetes.io/projected/b88ce1ac-1c8d-407d-8183-cdac9f354607-kube-api-access-89glt\") pod \"b88ce1ac-1c8d-407d-8183-cdac9f354607\" (UID: \"b88ce1ac-1c8d-407d-8183-cdac9f354607\") " Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.480723 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dllnv\" (UniqueName: \"kubernetes.io/projected/378ff2b6-e6b0-4cb0-a6ef-33e9ce1a7021-kube-api-access-dllnv\") pod \"378ff2b6-e6b0-4cb0-a6ef-33e9ce1a7021\" (UID: \"378ff2b6-e6b0-4cb0-a6ef-33e9ce1a7021\") " Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.480764 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxz47\" (UniqueName: \"kubernetes.io/projected/d3bd96cd-ca29-4a82-8c97-0709b76cd021-kube-api-access-fxz47\") pod \"d3bd96cd-ca29-4a82-8c97-0709b76cd021\" (UID: \"d3bd96cd-ca29-4a82-8c97-0709b76cd021\") " Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.480818 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3bd96cd-ca29-4a82-8c97-0709b76cd021-operator-scripts\") pod \"d3bd96cd-ca29-4a82-8c97-0709b76cd021\" (UID: \"d3bd96cd-ca29-4a82-8c97-0709b76cd021\") " Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.480872 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5b670080-8ca4-4e80-a1a6-ec7a63c951b6-operator-scripts\") pod \"5b670080-8ca4-4e80-a1a6-ec7a63c951b6\" (UID: \"5b670080-8ca4-4e80-a1a6-ec7a63c951b6\") " Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.480902 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b88ce1ac-1c8d-407d-8183-cdac9f354607-operator-scripts\") pod \"b88ce1ac-1c8d-407d-8183-cdac9f354607\" (UID: \"b88ce1ac-1c8d-407d-8183-cdac9f354607\") " Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.480939 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf5b318f-e3e1-45e0-b4d5-dd7fe656676f-operator-scripts\") pod \"cf5b318f-e3e1-45e0-b4d5-dd7fe656676f\" (UID: \"cf5b318f-e3e1-45e0-b4d5-dd7fe656676f\") " Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.481211 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/378ff2b6-e6b0-4cb0-a6ef-33e9ce1a7021-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "378ff2b6-e6b0-4cb0-a6ef-33e9ce1a7021" (UID: "378ff2b6-e6b0-4cb0-a6ef-33e9ce1a7021"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.481646 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b670080-8ca4-4e80-a1a6-ec7a63c951b6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5b670080-8ca4-4e80-a1a6-ec7a63c951b6" (UID: "5b670080-8ca4-4e80-a1a6-ec7a63c951b6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.481909 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3bd96cd-ca29-4a82-8c97-0709b76cd021-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d3bd96cd-ca29-4a82-8c97-0709b76cd021" (UID: "d3bd96cd-ca29-4a82-8c97-0709b76cd021"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.482519 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b88ce1ac-1c8d-407d-8183-cdac9f354607-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b88ce1ac-1c8d-407d-8183-cdac9f354607" (UID: "b88ce1ac-1c8d-407d-8183-cdac9f354607"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.484435 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf5b318f-e3e1-45e0-b4d5-dd7fe656676f-kube-api-access-rbrrl" (OuterVolumeSpecName: "kube-api-access-rbrrl") pod "cf5b318f-e3e1-45e0-b4d5-dd7fe656676f" (UID: "cf5b318f-e3e1-45e0-b4d5-dd7fe656676f"). InnerVolumeSpecName "kube-api-access-rbrrl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.485452 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b670080-8ca4-4e80-a1a6-ec7a63c951b6-kube-api-access-c74ct" (OuterVolumeSpecName: "kube-api-access-c74ct") pod "5b670080-8ca4-4e80-a1a6-ec7a63c951b6" (UID: "5b670080-8ca4-4e80-a1a6-ec7a63c951b6"). InnerVolumeSpecName "kube-api-access-c74ct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.487174 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/378ff2b6-e6b0-4cb0-a6ef-33e9ce1a7021-kube-api-access-dllnv" (OuterVolumeSpecName: "kube-api-access-dllnv") pod "378ff2b6-e6b0-4cb0-a6ef-33e9ce1a7021" (UID: "378ff2b6-e6b0-4cb0-a6ef-33e9ce1a7021"). InnerVolumeSpecName "kube-api-access-dllnv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.487947 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3bd96cd-ca29-4a82-8c97-0709b76cd021-kube-api-access-fxz47" (OuterVolumeSpecName: "kube-api-access-fxz47") pod "d3bd96cd-ca29-4a82-8c97-0709b76cd021" (UID: "d3bd96cd-ca29-4a82-8c97-0709b76cd021"). InnerVolumeSpecName "kube-api-access-fxz47". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.488084 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf5b318f-e3e1-45e0-b4d5-dd7fe656676f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cf5b318f-e3e1-45e0-b4d5-dd7fe656676f" (UID: "cf5b318f-e3e1-45e0-b4d5-dd7fe656676f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.488865 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf5b318f-e3e1-45e0-b4d5-dd7fe656676f-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.488895 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3452c4a6-ce88-4908-9ebf-6b53ccfcdc2e-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.488909 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qjffp\" (UniqueName: \"kubernetes.io/projected/3452c4a6-ce88-4908-9ebf-6b53ccfcdc2e-kube-api-access-qjffp\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.488924 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c74ct\" (UniqueName: \"kubernetes.io/projected/5b670080-8ca4-4e80-a1a6-ec7a63c951b6-kube-api-access-c74ct\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.488938 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbrrl\" (UniqueName: \"kubernetes.io/projected/cf5b318f-e3e1-45e0-b4d5-dd7fe656676f-kube-api-access-rbrrl\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.488951 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/378ff2b6-e6b0-4cb0-a6ef-33e9ce1a7021-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.488963 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dllnv\" (UniqueName: \"kubernetes.io/projected/378ff2b6-e6b0-4cb0-a6ef-33e9ce1a7021-kube-api-access-dllnv\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.488974 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxz47\" (UniqueName: \"kubernetes.io/projected/d3bd96cd-ca29-4a82-8c97-0709b76cd021-kube-api-access-fxz47\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.488984 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3bd96cd-ca29-4a82-8c97-0709b76cd021-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.488995 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5b670080-8ca4-4e80-a1a6-ec7a63c951b6-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.489007 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b88ce1ac-1c8d-407d-8183-cdac9f354607-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.489962 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b88ce1ac-1c8d-407d-8183-cdac9f354607-kube-api-access-89glt" (OuterVolumeSpecName: "kube-api-access-89glt") pod "b88ce1ac-1c8d-407d-8183-cdac9f354607" (UID: "b88ce1ac-1c8d-407d-8183-cdac9f354607"). InnerVolumeSpecName "kube-api-access-89glt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.590663 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-89glt\" (UniqueName: \"kubernetes.io/projected/b88ce1ac-1c8d-407d-8183-cdac9f354607-kube-api-access-89glt\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.678659 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f21cec8f-8cd4-4ecd-9800-3bb0c482fb66" path="/var/lib/kubelet/pods/f21cec8f-8cd4-4ecd-9800-3bb0c482fb66/volumes" Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.798295 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-6872-account-create-update-ndcpk" event={"ID":"d3bd96cd-ca29-4a82-8c97-0709b76cd021","Type":"ContainerDied","Data":"ab3103cfd6dde264bd6985d98d14271eb93b39b91b6a037feeed1a9c009d2589"} Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.798346 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab3103cfd6dde264bd6985d98d14271eb93b39b91b6a037feeed1a9c009d2589" Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.798307 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-6872-account-create-update-ndcpk" Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.804462 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-wq84l" Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.805340 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-wq84l" event={"ID":"5b670080-8ca4-4e80-a1a6-ec7a63c951b6","Type":"ContainerDied","Data":"60a83bc58f9bddb9bf5d9e9acbe836faadef07dabfc04886d2b6bd78d589ad25"} Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.805395 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="60a83bc58f9bddb9bf5d9e9acbe836faadef07dabfc04886d2b6bd78d589ad25" Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.810873 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-zcnn2" event={"ID":"378ff2b6-e6b0-4cb0-a6ef-33e9ce1a7021","Type":"ContainerDied","Data":"9bf1a61ddf57c2567f3638d3f5eab042c6c76a057190d6721b3824551befc6c0"} Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.810942 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9bf1a61ddf57c2567f3638d3f5eab042c6c76a057190d6721b3824551befc6c0" Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.811011 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-zcnn2" Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.816242 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-hn8xp" event={"ID":"cf5b318f-e3e1-45e0-b4d5-dd7fe656676f","Type":"ContainerDied","Data":"45854e1296321401f23fca00a556958adcaf2c47d25ba35f94c0614e0ecde256"} Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.816294 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="45854e1296321401f23fca00a556958adcaf2c47d25ba35f94c0614e0ecde256" Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.816353 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-hn8xp" Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.821597 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-da5a-account-create-update-ntzr7" event={"ID":"b88ce1ac-1c8d-407d-8183-cdac9f354607","Type":"ContainerDied","Data":"aa4dd2af37bb6eb1b45670f388f98036191d2ca766e280c90fdfd4fdf121f0cf"} Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.821638 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa4dd2af37bb6eb1b45670f388f98036191d2ca766e280c90fdfd4fdf121f0cf" Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.821715 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-da5a-account-create-update-ntzr7" Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.826736 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-48c8-account-create-update-t6p84" event={"ID":"3452c4a6-ce88-4908-9ebf-6b53ccfcdc2e","Type":"ContainerDied","Data":"da5fd9e5487f0a4446471475dee93e2a83d70dbacdaac9e9a4231d2285bb023c"} Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.828217 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="da5fd9e5487f0a4446471475dee93e2a83d70dbacdaac9e9a4231d2285bb023c" Nov 29 04:56:30 crc kubenswrapper[4799]: I1129 04:56:30.827478 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-48c8-account-create-update-t6p84" Nov 29 04:56:31 crc kubenswrapper[4799]: I1129 04:56:31.846755 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e97e3977-fa5b-4723-9dcc-55907243890d","Type":"ContainerStarted","Data":"eeaceac970c3ad6a835b7e8a9e279553ea5515392f0078fd0b6e54e98f247ad1"} Nov 29 04:56:31 crc kubenswrapper[4799]: I1129 04:56:31.847212 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e97e3977-fa5b-4723-9dcc-55907243890d" containerName="ceilometer-central-agent" containerID="cri-o://f7817ff369dcbebba92bf72b341244eae9027d4b2c5da4c4c22a39f691bb3040" gracePeriod=30 Nov 29 04:56:31 crc kubenswrapper[4799]: I1129 04:56:31.847250 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e97e3977-fa5b-4723-9dcc-55907243890d" containerName="proxy-httpd" containerID="cri-o://eeaceac970c3ad6a835b7e8a9e279553ea5515392f0078fd0b6e54e98f247ad1" gracePeriod=30 Nov 29 04:56:31 crc kubenswrapper[4799]: I1129 04:56:31.847291 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 29 04:56:31 crc kubenswrapper[4799]: I1129 04:56:31.847364 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e97e3977-fa5b-4723-9dcc-55907243890d" containerName="ceilometer-notification-agent" containerID="cri-o://517be1f3d946d604be5a9dd7f7b645f08b271f832012ff40acbbab0df72a7702" gracePeriod=30 Nov 29 04:56:31 crc kubenswrapper[4799]: I1129 04:56:31.847354 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e97e3977-fa5b-4723-9dcc-55907243890d" containerName="sg-core" containerID="cri-o://7ad491dbe6e90a0e0aeeb6bcba1bbe8615e3a298443179011c8544976f56b561" gracePeriod=30 Nov 29 04:56:31 crc kubenswrapper[4799]: I1129 04:56:31.884652 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.842343736 podStartE2EDuration="7.884624396s" podCreationTimestamp="2025-11-29 04:56:24 +0000 UTC" firstStartedPulling="2025-11-29 04:56:25.753726292 +0000 UTC m=+1061.396656692" lastFinishedPulling="2025-11-29 04:56:30.796006952 +0000 UTC m=+1066.438937352" observedRunningTime="2025-11-29 04:56:31.870808848 +0000 UTC m=+1067.513739278" watchObservedRunningTime="2025-11-29 04:56:31.884624396 +0000 UTC m=+1067.527554816" Nov 29 04:56:32 crc kubenswrapper[4799]: I1129 04:56:32.857125 4799 generic.go:334] "Generic (PLEG): container finished" podID="e97e3977-fa5b-4723-9dcc-55907243890d" containerID="eeaceac970c3ad6a835b7e8a9e279553ea5515392f0078fd0b6e54e98f247ad1" exitCode=0 Nov 29 04:56:32 crc kubenswrapper[4799]: I1129 04:56:32.857162 4799 generic.go:334] "Generic (PLEG): container finished" podID="e97e3977-fa5b-4723-9dcc-55907243890d" containerID="7ad491dbe6e90a0e0aeeb6bcba1bbe8615e3a298443179011c8544976f56b561" exitCode=2 Nov 29 04:56:32 crc kubenswrapper[4799]: I1129 04:56:32.857173 4799 generic.go:334] "Generic (PLEG): container finished" podID="e97e3977-fa5b-4723-9dcc-55907243890d" containerID="517be1f3d946d604be5a9dd7f7b645f08b271f832012ff40acbbab0df72a7702" exitCode=0 Nov 29 04:56:32 crc kubenswrapper[4799]: I1129 04:56:32.857195 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e97e3977-fa5b-4723-9dcc-55907243890d","Type":"ContainerDied","Data":"eeaceac970c3ad6a835b7e8a9e279553ea5515392f0078fd0b6e54e98f247ad1"} Nov 29 04:56:32 crc kubenswrapper[4799]: I1129 04:56:32.857226 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e97e3977-fa5b-4723-9dcc-55907243890d","Type":"ContainerDied","Data":"7ad491dbe6e90a0e0aeeb6bcba1bbe8615e3a298443179011c8544976f56b561"} Nov 29 04:56:32 crc kubenswrapper[4799]: I1129 04:56:32.857238 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e97e3977-fa5b-4723-9dcc-55907243890d","Type":"ContainerDied","Data":"517be1f3d946d604be5a9dd7f7b645f08b271f832012ff40acbbab0df72a7702"} Nov 29 04:56:34 crc kubenswrapper[4799]: I1129 04:56:34.608344 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-5g6rp"] Nov 29 04:56:34 crc kubenswrapper[4799]: E1129 04:56:34.609189 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f21cec8f-8cd4-4ecd-9800-3bb0c482fb66" containerName="neutron-httpd" Nov 29 04:56:34 crc kubenswrapper[4799]: I1129 04:56:34.609208 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f21cec8f-8cd4-4ecd-9800-3bb0c482fb66" containerName="neutron-httpd" Nov 29 04:56:34 crc kubenswrapper[4799]: E1129 04:56:34.609229 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3452c4a6-ce88-4908-9ebf-6b53ccfcdc2e" containerName="mariadb-account-create-update" Nov 29 04:56:34 crc kubenswrapper[4799]: I1129 04:56:34.609236 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="3452c4a6-ce88-4908-9ebf-6b53ccfcdc2e" containerName="mariadb-account-create-update" Nov 29 04:56:34 crc kubenswrapper[4799]: E1129 04:56:34.609255 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="378ff2b6-e6b0-4cb0-a6ef-33e9ce1a7021" containerName="mariadb-database-create" Nov 29 04:56:34 crc kubenswrapper[4799]: I1129 04:56:34.609276 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="378ff2b6-e6b0-4cb0-a6ef-33e9ce1a7021" containerName="mariadb-database-create" Nov 29 04:56:34 crc kubenswrapper[4799]: E1129 04:56:34.609289 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b670080-8ca4-4e80-a1a6-ec7a63c951b6" containerName="mariadb-database-create" Nov 29 04:56:34 crc kubenswrapper[4799]: I1129 04:56:34.609296 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b670080-8ca4-4e80-a1a6-ec7a63c951b6" containerName="mariadb-database-create" Nov 29 04:56:34 crc kubenswrapper[4799]: E1129 04:56:34.609305 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3bd96cd-ca29-4a82-8c97-0709b76cd021" containerName="mariadb-account-create-update" Nov 29 04:56:34 crc kubenswrapper[4799]: I1129 04:56:34.609312 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3bd96cd-ca29-4a82-8c97-0709b76cd021" containerName="mariadb-account-create-update" Nov 29 04:56:34 crc kubenswrapper[4799]: E1129 04:56:34.609320 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf5b318f-e3e1-45e0-b4d5-dd7fe656676f" containerName="mariadb-database-create" Nov 29 04:56:34 crc kubenswrapper[4799]: I1129 04:56:34.609327 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf5b318f-e3e1-45e0-b4d5-dd7fe656676f" containerName="mariadb-database-create" Nov 29 04:56:34 crc kubenswrapper[4799]: E1129 04:56:34.609353 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f21cec8f-8cd4-4ecd-9800-3bb0c482fb66" containerName="neutron-api" Nov 29 04:56:34 crc kubenswrapper[4799]: I1129 04:56:34.609362 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f21cec8f-8cd4-4ecd-9800-3bb0c482fb66" containerName="neutron-api" Nov 29 04:56:34 crc kubenswrapper[4799]: E1129 04:56:34.609371 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b88ce1ac-1c8d-407d-8183-cdac9f354607" containerName="mariadb-account-create-update" Nov 29 04:56:34 crc kubenswrapper[4799]: I1129 04:56:34.609378 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="b88ce1ac-1c8d-407d-8183-cdac9f354607" containerName="mariadb-account-create-update" Nov 29 04:56:34 crc kubenswrapper[4799]: I1129 04:56:34.609572 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf5b318f-e3e1-45e0-b4d5-dd7fe656676f" containerName="mariadb-database-create" Nov 29 04:56:34 crc kubenswrapper[4799]: I1129 04:56:34.609588 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="378ff2b6-e6b0-4cb0-a6ef-33e9ce1a7021" containerName="mariadb-database-create" Nov 29 04:56:34 crc kubenswrapper[4799]: I1129 04:56:34.609596 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="f21cec8f-8cd4-4ecd-9800-3bb0c482fb66" containerName="neutron-httpd" Nov 29 04:56:34 crc kubenswrapper[4799]: I1129 04:56:34.609605 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="3452c4a6-ce88-4908-9ebf-6b53ccfcdc2e" containerName="mariadb-account-create-update" Nov 29 04:56:34 crc kubenswrapper[4799]: I1129 04:56:34.609617 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="f21cec8f-8cd4-4ecd-9800-3bb0c482fb66" containerName="neutron-api" Nov 29 04:56:34 crc kubenswrapper[4799]: I1129 04:56:34.609631 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b670080-8ca4-4e80-a1a6-ec7a63c951b6" containerName="mariadb-database-create" Nov 29 04:56:34 crc kubenswrapper[4799]: I1129 04:56:34.609642 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3bd96cd-ca29-4a82-8c97-0709b76cd021" containerName="mariadb-account-create-update" Nov 29 04:56:34 crc kubenswrapper[4799]: I1129 04:56:34.609654 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="b88ce1ac-1c8d-407d-8183-cdac9f354607" containerName="mariadb-account-create-update" Nov 29 04:56:34 crc kubenswrapper[4799]: I1129 04:56:34.610367 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-5g6rp" Nov 29 04:56:34 crc kubenswrapper[4799]: I1129 04:56:34.612513 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Nov 29 04:56:34 crc kubenswrapper[4799]: I1129 04:56:34.613064 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-5hbsz" Nov 29 04:56:34 crc kubenswrapper[4799]: I1129 04:56:34.614647 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 29 04:56:34 crc kubenswrapper[4799]: I1129 04:56:34.632921 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-5g6rp"] Nov 29 04:56:34 crc kubenswrapper[4799]: I1129 04:56:34.815699 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f53c1df-8c43-4a5a-94c7-7d48fedf34ac-config-data\") pod \"nova-cell0-conductor-db-sync-5g6rp\" (UID: \"6f53c1df-8c43-4a5a-94c7-7d48fedf34ac\") " pod="openstack/nova-cell0-conductor-db-sync-5g6rp" Nov 29 04:56:34 crc kubenswrapper[4799]: I1129 04:56:34.815782 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5l2ql\" (UniqueName: \"kubernetes.io/projected/6f53c1df-8c43-4a5a-94c7-7d48fedf34ac-kube-api-access-5l2ql\") pod \"nova-cell0-conductor-db-sync-5g6rp\" (UID: \"6f53c1df-8c43-4a5a-94c7-7d48fedf34ac\") " pod="openstack/nova-cell0-conductor-db-sync-5g6rp" Nov 29 04:56:34 crc kubenswrapper[4799]: I1129 04:56:34.815903 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f53c1df-8c43-4a5a-94c7-7d48fedf34ac-scripts\") pod \"nova-cell0-conductor-db-sync-5g6rp\" (UID: \"6f53c1df-8c43-4a5a-94c7-7d48fedf34ac\") " pod="openstack/nova-cell0-conductor-db-sync-5g6rp" Nov 29 04:56:34 crc kubenswrapper[4799]: I1129 04:56:34.815931 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f53c1df-8c43-4a5a-94c7-7d48fedf34ac-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-5g6rp\" (UID: \"6f53c1df-8c43-4a5a-94c7-7d48fedf34ac\") " pod="openstack/nova-cell0-conductor-db-sync-5g6rp" Nov 29 04:56:34 crc kubenswrapper[4799]: I1129 04:56:34.917352 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f53c1df-8c43-4a5a-94c7-7d48fedf34ac-scripts\") pod \"nova-cell0-conductor-db-sync-5g6rp\" (UID: \"6f53c1df-8c43-4a5a-94c7-7d48fedf34ac\") " pod="openstack/nova-cell0-conductor-db-sync-5g6rp" Nov 29 04:56:34 crc kubenswrapper[4799]: I1129 04:56:34.917779 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f53c1df-8c43-4a5a-94c7-7d48fedf34ac-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-5g6rp\" (UID: \"6f53c1df-8c43-4a5a-94c7-7d48fedf34ac\") " pod="openstack/nova-cell0-conductor-db-sync-5g6rp" Nov 29 04:56:34 crc kubenswrapper[4799]: I1129 04:56:34.917922 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f53c1df-8c43-4a5a-94c7-7d48fedf34ac-config-data\") pod \"nova-cell0-conductor-db-sync-5g6rp\" (UID: \"6f53c1df-8c43-4a5a-94c7-7d48fedf34ac\") " pod="openstack/nova-cell0-conductor-db-sync-5g6rp" Nov 29 04:56:34 crc kubenswrapper[4799]: I1129 04:56:34.917988 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5l2ql\" (UniqueName: \"kubernetes.io/projected/6f53c1df-8c43-4a5a-94c7-7d48fedf34ac-kube-api-access-5l2ql\") pod \"nova-cell0-conductor-db-sync-5g6rp\" (UID: \"6f53c1df-8c43-4a5a-94c7-7d48fedf34ac\") " pod="openstack/nova-cell0-conductor-db-sync-5g6rp" Nov 29 04:56:34 crc kubenswrapper[4799]: I1129 04:56:34.932194 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f53c1df-8c43-4a5a-94c7-7d48fedf34ac-scripts\") pod \"nova-cell0-conductor-db-sync-5g6rp\" (UID: \"6f53c1df-8c43-4a5a-94c7-7d48fedf34ac\") " pod="openstack/nova-cell0-conductor-db-sync-5g6rp" Nov 29 04:56:34 crc kubenswrapper[4799]: I1129 04:56:34.932434 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f53c1df-8c43-4a5a-94c7-7d48fedf34ac-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-5g6rp\" (UID: \"6f53c1df-8c43-4a5a-94c7-7d48fedf34ac\") " pod="openstack/nova-cell0-conductor-db-sync-5g6rp" Nov 29 04:56:34 crc kubenswrapper[4799]: I1129 04:56:34.932728 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f53c1df-8c43-4a5a-94c7-7d48fedf34ac-config-data\") pod \"nova-cell0-conductor-db-sync-5g6rp\" (UID: \"6f53c1df-8c43-4a5a-94c7-7d48fedf34ac\") " pod="openstack/nova-cell0-conductor-db-sync-5g6rp" Nov 29 04:56:34 crc kubenswrapper[4799]: I1129 04:56:34.937203 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5l2ql\" (UniqueName: \"kubernetes.io/projected/6f53c1df-8c43-4a5a-94c7-7d48fedf34ac-kube-api-access-5l2ql\") pod \"nova-cell0-conductor-db-sync-5g6rp\" (UID: \"6f53c1df-8c43-4a5a-94c7-7d48fedf34ac\") " pod="openstack/nova-cell0-conductor-db-sync-5g6rp" Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.037221 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-5g6rp" Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.489830 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-5g6rp"] Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.541500 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.633953 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e97e3977-fa5b-4723-9dcc-55907243890d-sg-core-conf-yaml\") pod \"e97e3977-fa5b-4723-9dcc-55907243890d\" (UID: \"e97e3977-fa5b-4723-9dcc-55907243890d\") " Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.635025 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e97e3977-fa5b-4723-9dcc-55907243890d-config-data\") pod \"e97e3977-fa5b-4723-9dcc-55907243890d\" (UID: \"e97e3977-fa5b-4723-9dcc-55907243890d\") " Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.635159 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e97e3977-fa5b-4723-9dcc-55907243890d-log-httpd\") pod \"e97e3977-fa5b-4723-9dcc-55907243890d\" (UID: \"e97e3977-fa5b-4723-9dcc-55907243890d\") " Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.635284 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e97e3977-fa5b-4723-9dcc-55907243890d-combined-ca-bundle\") pod \"e97e3977-fa5b-4723-9dcc-55907243890d\" (UID: \"e97e3977-fa5b-4723-9dcc-55907243890d\") " Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.635476 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fn2td\" (UniqueName: \"kubernetes.io/projected/e97e3977-fa5b-4723-9dcc-55907243890d-kube-api-access-fn2td\") pod \"e97e3977-fa5b-4723-9dcc-55907243890d\" (UID: \"e97e3977-fa5b-4723-9dcc-55907243890d\") " Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.635579 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e97e3977-fa5b-4723-9dcc-55907243890d-scripts\") pod \"e97e3977-fa5b-4723-9dcc-55907243890d\" (UID: \"e97e3977-fa5b-4723-9dcc-55907243890d\") " Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.635687 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e97e3977-fa5b-4723-9dcc-55907243890d-run-httpd\") pod \"e97e3977-fa5b-4723-9dcc-55907243890d\" (UID: \"e97e3977-fa5b-4723-9dcc-55907243890d\") " Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.635892 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e97e3977-fa5b-4723-9dcc-55907243890d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e97e3977-fa5b-4723-9dcc-55907243890d" (UID: "e97e3977-fa5b-4723-9dcc-55907243890d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.636139 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e97e3977-fa5b-4723-9dcc-55907243890d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e97e3977-fa5b-4723-9dcc-55907243890d" (UID: "e97e3977-fa5b-4723-9dcc-55907243890d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.636517 4799 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e97e3977-fa5b-4723-9dcc-55907243890d-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.636627 4799 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e97e3977-fa5b-4723-9dcc-55907243890d-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.639600 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e97e3977-fa5b-4723-9dcc-55907243890d-scripts" (OuterVolumeSpecName: "scripts") pod "e97e3977-fa5b-4723-9dcc-55907243890d" (UID: "e97e3977-fa5b-4723-9dcc-55907243890d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.640830 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e97e3977-fa5b-4723-9dcc-55907243890d-kube-api-access-fn2td" (OuterVolumeSpecName: "kube-api-access-fn2td") pod "e97e3977-fa5b-4723-9dcc-55907243890d" (UID: "e97e3977-fa5b-4723-9dcc-55907243890d"). InnerVolumeSpecName "kube-api-access-fn2td". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.673034 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e97e3977-fa5b-4723-9dcc-55907243890d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e97e3977-fa5b-4723-9dcc-55907243890d" (UID: "e97e3977-fa5b-4723-9dcc-55907243890d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.709328 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e97e3977-fa5b-4723-9dcc-55907243890d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e97e3977-fa5b-4723-9dcc-55907243890d" (UID: "e97e3977-fa5b-4723-9dcc-55907243890d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.734042 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e97e3977-fa5b-4723-9dcc-55907243890d-config-data" (OuterVolumeSpecName: "config-data") pod "e97e3977-fa5b-4723-9dcc-55907243890d" (UID: "e97e3977-fa5b-4723-9dcc-55907243890d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.738232 4799 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e97e3977-fa5b-4723-9dcc-55907243890d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.738265 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e97e3977-fa5b-4723-9dcc-55907243890d-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.738278 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e97e3977-fa5b-4723-9dcc-55907243890d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.738289 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fn2td\" (UniqueName: \"kubernetes.io/projected/e97e3977-fa5b-4723-9dcc-55907243890d-kube-api-access-fn2td\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.738303 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e97e3977-fa5b-4723-9dcc-55907243890d-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.882557 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-5g6rp" event={"ID":"6f53c1df-8c43-4a5a-94c7-7d48fedf34ac","Type":"ContainerStarted","Data":"330d4eca3cb2178385d83fbf34af6bb57705209a177a56249e53527480d404da"} Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.884782 4799 generic.go:334] "Generic (PLEG): container finished" podID="e97e3977-fa5b-4723-9dcc-55907243890d" containerID="f7817ff369dcbebba92bf72b341244eae9027d4b2c5da4c4c22a39f691bb3040" exitCode=0 Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.884849 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e97e3977-fa5b-4723-9dcc-55907243890d","Type":"ContainerDied","Data":"f7817ff369dcbebba92bf72b341244eae9027d4b2c5da4c4c22a39f691bb3040"} Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.884884 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e97e3977-fa5b-4723-9dcc-55907243890d","Type":"ContainerDied","Data":"2b99e42389aa495845f6d8c05f71ea3208c64ae83ba7bb4fc0f4909d39a4e449"} Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.884907 4799 scope.go:117] "RemoveContainer" containerID="eeaceac970c3ad6a835b7e8a9e279553ea5515392f0078fd0b6e54e98f247ad1" Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.885103 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.908149 4799 scope.go:117] "RemoveContainer" containerID="7ad491dbe6e90a0e0aeeb6bcba1bbe8615e3a298443179011c8544976f56b561" Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.930086 4799 scope.go:117] "RemoveContainer" containerID="517be1f3d946d604be5a9dd7f7b645f08b271f832012ff40acbbab0df72a7702" Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.939672 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.949122 4799 scope.go:117] "RemoveContainer" containerID="f7817ff369dcbebba92bf72b341244eae9027d4b2c5da4c4c22a39f691bb3040" Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.950239 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.968936 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 29 04:56:35 crc kubenswrapper[4799]: E1129 04:56:35.969485 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e97e3977-fa5b-4723-9dcc-55907243890d" containerName="sg-core" Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.969503 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e97e3977-fa5b-4723-9dcc-55907243890d" containerName="sg-core" Nov 29 04:56:35 crc kubenswrapper[4799]: E1129 04:56:35.969532 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e97e3977-fa5b-4723-9dcc-55907243890d" containerName="ceilometer-central-agent" Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.969541 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e97e3977-fa5b-4723-9dcc-55907243890d" containerName="ceilometer-central-agent" Nov 29 04:56:35 crc kubenswrapper[4799]: E1129 04:56:35.969556 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e97e3977-fa5b-4723-9dcc-55907243890d" containerName="proxy-httpd" Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.969563 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e97e3977-fa5b-4723-9dcc-55907243890d" containerName="proxy-httpd" Nov 29 04:56:35 crc kubenswrapper[4799]: E1129 04:56:35.969574 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e97e3977-fa5b-4723-9dcc-55907243890d" containerName="ceilometer-notification-agent" Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.969582 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e97e3977-fa5b-4723-9dcc-55907243890d" containerName="ceilometer-notification-agent" Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.969869 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="e97e3977-fa5b-4723-9dcc-55907243890d" containerName="sg-core" Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.969884 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="e97e3977-fa5b-4723-9dcc-55907243890d" containerName="ceilometer-notification-agent" Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.969902 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="e97e3977-fa5b-4723-9dcc-55907243890d" containerName="ceilometer-central-agent" Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.969922 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="e97e3977-fa5b-4723-9dcc-55907243890d" containerName="proxy-httpd" Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.972341 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.978622 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.979317 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.982552 4799 scope.go:117] "RemoveContainer" containerID="eeaceac970c3ad6a835b7e8a9e279553ea5515392f0078fd0b6e54e98f247ad1" Nov 29 04:56:35 crc kubenswrapper[4799]: E1129 04:56:35.986194 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eeaceac970c3ad6a835b7e8a9e279553ea5515392f0078fd0b6e54e98f247ad1\": container with ID starting with eeaceac970c3ad6a835b7e8a9e279553ea5515392f0078fd0b6e54e98f247ad1 not found: ID does not exist" containerID="eeaceac970c3ad6a835b7e8a9e279553ea5515392f0078fd0b6e54e98f247ad1" Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.986233 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eeaceac970c3ad6a835b7e8a9e279553ea5515392f0078fd0b6e54e98f247ad1"} err="failed to get container status \"eeaceac970c3ad6a835b7e8a9e279553ea5515392f0078fd0b6e54e98f247ad1\": rpc error: code = NotFound desc = could not find container \"eeaceac970c3ad6a835b7e8a9e279553ea5515392f0078fd0b6e54e98f247ad1\": container with ID starting with eeaceac970c3ad6a835b7e8a9e279553ea5515392f0078fd0b6e54e98f247ad1 not found: ID does not exist" Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.986258 4799 scope.go:117] "RemoveContainer" containerID="7ad491dbe6e90a0e0aeeb6bcba1bbe8615e3a298443179011c8544976f56b561" Nov 29 04:56:35 crc kubenswrapper[4799]: E1129 04:56:35.989327 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ad491dbe6e90a0e0aeeb6bcba1bbe8615e3a298443179011c8544976f56b561\": container with ID starting with 7ad491dbe6e90a0e0aeeb6bcba1bbe8615e3a298443179011c8544976f56b561 not found: ID does not exist" containerID="7ad491dbe6e90a0e0aeeb6bcba1bbe8615e3a298443179011c8544976f56b561" Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.989389 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ad491dbe6e90a0e0aeeb6bcba1bbe8615e3a298443179011c8544976f56b561"} err="failed to get container status \"7ad491dbe6e90a0e0aeeb6bcba1bbe8615e3a298443179011c8544976f56b561\": rpc error: code = NotFound desc = could not find container \"7ad491dbe6e90a0e0aeeb6bcba1bbe8615e3a298443179011c8544976f56b561\": container with ID starting with 7ad491dbe6e90a0e0aeeb6bcba1bbe8615e3a298443179011c8544976f56b561 not found: ID does not exist" Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.989423 4799 scope.go:117] "RemoveContainer" containerID="517be1f3d946d604be5a9dd7f7b645f08b271f832012ff40acbbab0df72a7702" Nov 29 04:56:35 crc kubenswrapper[4799]: E1129 04:56:35.992874 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"517be1f3d946d604be5a9dd7f7b645f08b271f832012ff40acbbab0df72a7702\": container with ID starting with 517be1f3d946d604be5a9dd7f7b645f08b271f832012ff40acbbab0df72a7702 not found: ID does not exist" containerID="517be1f3d946d604be5a9dd7f7b645f08b271f832012ff40acbbab0df72a7702" Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.992906 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"517be1f3d946d604be5a9dd7f7b645f08b271f832012ff40acbbab0df72a7702"} err="failed to get container status \"517be1f3d946d604be5a9dd7f7b645f08b271f832012ff40acbbab0df72a7702\": rpc error: code = NotFound desc = could not find container \"517be1f3d946d604be5a9dd7f7b645f08b271f832012ff40acbbab0df72a7702\": container with ID starting with 517be1f3d946d604be5a9dd7f7b645f08b271f832012ff40acbbab0df72a7702 not found: ID does not exist" Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.992923 4799 scope.go:117] "RemoveContainer" containerID="f7817ff369dcbebba92bf72b341244eae9027d4b2c5da4c4c22a39f691bb3040" Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.997270 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 29 04:56:35 crc kubenswrapper[4799]: E1129 04:56:35.999365 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7817ff369dcbebba92bf72b341244eae9027d4b2c5da4c4c22a39f691bb3040\": container with ID starting with f7817ff369dcbebba92bf72b341244eae9027d4b2c5da4c4c22a39f691bb3040 not found: ID does not exist" containerID="f7817ff369dcbebba92bf72b341244eae9027d4b2c5da4c4c22a39f691bb3040" Nov 29 04:56:35 crc kubenswrapper[4799]: I1129 04:56:35.999427 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7817ff369dcbebba92bf72b341244eae9027d4b2c5da4c4c22a39f691bb3040"} err="failed to get container status \"f7817ff369dcbebba92bf72b341244eae9027d4b2c5da4c4c22a39f691bb3040\": rpc error: code = NotFound desc = could not find container \"f7817ff369dcbebba92bf72b341244eae9027d4b2c5da4c4c22a39f691bb3040\": container with ID starting with f7817ff369dcbebba92bf72b341244eae9027d4b2c5da4c4c22a39f691bb3040 not found: ID does not exist" Nov 29 04:56:36 crc kubenswrapper[4799]: I1129 04:56:36.044341 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4bba1b4-2512-4ced-9c5f-43567fd3535b-scripts\") pod \"ceilometer-0\" (UID: \"c4bba1b4-2512-4ced-9c5f-43567fd3535b\") " pod="openstack/ceilometer-0" Nov 29 04:56:36 crc kubenswrapper[4799]: I1129 04:56:36.044445 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4bba1b4-2512-4ced-9c5f-43567fd3535b-log-httpd\") pod \"ceilometer-0\" (UID: \"c4bba1b4-2512-4ced-9c5f-43567fd3535b\") " pod="openstack/ceilometer-0" Nov 29 04:56:36 crc kubenswrapper[4799]: I1129 04:56:36.044539 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62v52\" (UniqueName: \"kubernetes.io/projected/c4bba1b4-2512-4ced-9c5f-43567fd3535b-kube-api-access-62v52\") pod \"ceilometer-0\" (UID: \"c4bba1b4-2512-4ced-9c5f-43567fd3535b\") " pod="openstack/ceilometer-0" Nov 29 04:56:36 crc kubenswrapper[4799]: I1129 04:56:36.044609 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4bba1b4-2512-4ced-9c5f-43567fd3535b-config-data\") pod \"ceilometer-0\" (UID: \"c4bba1b4-2512-4ced-9c5f-43567fd3535b\") " pod="openstack/ceilometer-0" Nov 29 04:56:36 crc kubenswrapper[4799]: I1129 04:56:36.044647 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4bba1b4-2512-4ced-9c5f-43567fd3535b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c4bba1b4-2512-4ced-9c5f-43567fd3535b\") " pod="openstack/ceilometer-0" Nov 29 04:56:36 crc kubenswrapper[4799]: I1129 04:56:36.044674 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c4bba1b4-2512-4ced-9c5f-43567fd3535b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c4bba1b4-2512-4ced-9c5f-43567fd3535b\") " pod="openstack/ceilometer-0" Nov 29 04:56:36 crc kubenswrapper[4799]: I1129 04:56:36.044731 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4bba1b4-2512-4ced-9c5f-43567fd3535b-run-httpd\") pod \"ceilometer-0\" (UID: \"c4bba1b4-2512-4ced-9c5f-43567fd3535b\") " pod="openstack/ceilometer-0" Nov 29 04:56:36 crc kubenswrapper[4799]: I1129 04:56:36.146832 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62v52\" (UniqueName: \"kubernetes.io/projected/c4bba1b4-2512-4ced-9c5f-43567fd3535b-kube-api-access-62v52\") pod \"ceilometer-0\" (UID: \"c4bba1b4-2512-4ced-9c5f-43567fd3535b\") " pod="openstack/ceilometer-0" Nov 29 04:56:36 crc kubenswrapper[4799]: I1129 04:56:36.146933 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4bba1b4-2512-4ced-9c5f-43567fd3535b-config-data\") pod \"ceilometer-0\" (UID: \"c4bba1b4-2512-4ced-9c5f-43567fd3535b\") " pod="openstack/ceilometer-0" Nov 29 04:56:36 crc kubenswrapper[4799]: I1129 04:56:36.146972 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4bba1b4-2512-4ced-9c5f-43567fd3535b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c4bba1b4-2512-4ced-9c5f-43567fd3535b\") " pod="openstack/ceilometer-0" Nov 29 04:56:36 crc kubenswrapper[4799]: I1129 04:56:36.147000 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c4bba1b4-2512-4ced-9c5f-43567fd3535b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c4bba1b4-2512-4ced-9c5f-43567fd3535b\") " pod="openstack/ceilometer-0" Nov 29 04:56:36 crc kubenswrapper[4799]: I1129 04:56:36.147064 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4bba1b4-2512-4ced-9c5f-43567fd3535b-run-httpd\") pod \"ceilometer-0\" (UID: \"c4bba1b4-2512-4ced-9c5f-43567fd3535b\") " pod="openstack/ceilometer-0" Nov 29 04:56:36 crc kubenswrapper[4799]: I1129 04:56:36.147139 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4bba1b4-2512-4ced-9c5f-43567fd3535b-scripts\") pod \"ceilometer-0\" (UID: \"c4bba1b4-2512-4ced-9c5f-43567fd3535b\") " pod="openstack/ceilometer-0" Nov 29 04:56:36 crc kubenswrapper[4799]: I1129 04:56:36.147185 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4bba1b4-2512-4ced-9c5f-43567fd3535b-log-httpd\") pod \"ceilometer-0\" (UID: \"c4bba1b4-2512-4ced-9c5f-43567fd3535b\") " pod="openstack/ceilometer-0" Nov 29 04:56:36 crc kubenswrapper[4799]: I1129 04:56:36.148101 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4bba1b4-2512-4ced-9c5f-43567fd3535b-run-httpd\") pod \"ceilometer-0\" (UID: \"c4bba1b4-2512-4ced-9c5f-43567fd3535b\") " pod="openstack/ceilometer-0" Nov 29 04:56:36 crc kubenswrapper[4799]: I1129 04:56:36.151075 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4bba1b4-2512-4ced-9c5f-43567fd3535b-log-httpd\") pod \"ceilometer-0\" (UID: \"c4bba1b4-2512-4ced-9c5f-43567fd3535b\") " pod="openstack/ceilometer-0" Nov 29 04:56:36 crc kubenswrapper[4799]: I1129 04:56:36.154091 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4bba1b4-2512-4ced-9c5f-43567fd3535b-config-data\") pod \"ceilometer-0\" (UID: \"c4bba1b4-2512-4ced-9c5f-43567fd3535b\") " pod="openstack/ceilometer-0" Nov 29 04:56:36 crc kubenswrapper[4799]: I1129 04:56:36.155385 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4bba1b4-2512-4ced-9c5f-43567fd3535b-scripts\") pod \"ceilometer-0\" (UID: \"c4bba1b4-2512-4ced-9c5f-43567fd3535b\") " pod="openstack/ceilometer-0" Nov 29 04:56:36 crc kubenswrapper[4799]: I1129 04:56:36.162537 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4bba1b4-2512-4ced-9c5f-43567fd3535b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c4bba1b4-2512-4ced-9c5f-43567fd3535b\") " pod="openstack/ceilometer-0" Nov 29 04:56:36 crc kubenswrapper[4799]: I1129 04:56:36.164108 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62v52\" (UniqueName: \"kubernetes.io/projected/c4bba1b4-2512-4ced-9c5f-43567fd3535b-kube-api-access-62v52\") pod \"ceilometer-0\" (UID: \"c4bba1b4-2512-4ced-9c5f-43567fd3535b\") " pod="openstack/ceilometer-0" Nov 29 04:56:36 crc kubenswrapper[4799]: I1129 04:56:36.165465 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c4bba1b4-2512-4ced-9c5f-43567fd3535b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c4bba1b4-2512-4ced-9c5f-43567fd3535b\") " pod="openstack/ceilometer-0" Nov 29 04:56:36 crc kubenswrapper[4799]: I1129 04:56:36.316250 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 04:56:36 crc kubenswrapper[4799]: I1129 04:56:36.674655 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e97e3977-fa5b-4723-9dcc-55907243890d" path="/var/lib/kubelet/pods/e97e3977-fa5b-4723-9dcc-55907243890d/volumes" Nov 29 04:56:36 crc kubenswrapper[4799]: W1129 04:56:36.958912 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc4bba1b4_2512_4ced_9c5f_43567fd3535b.slice/crio-b61227d6a655050ee38dc18d24bb88c8f84832b215c9dca2b78b4391c4e52ed8 WatchSource:0}: Error finding container b61227d6a655050ee38dc18d24bb88c8f84832b215c9dca2b78b4391c4e52ed8: Status 404 returned error can't find the container with id b61227d6a655050ee38dc18d24bb88c8f84832b215c9dca2b78b4391c4e52ed8 Nov 29 04:56:36 crc kubenswrapper[4799]: I1129 04:56:36.961246 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 29 04:56:37 crc kubenswrapper[4799]: I1129 04:56:37.937059 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4bba1b4-2512-4ced-9c5f-43567fd3535b","Type":"ContainerStarted","Data":"b61227d6a655050ee38dc18d24bb88c8f84832b215c9dca2b78b4391c4e52ed8"} Nov 29 04:56:38 crc kubenswrapper[4799]: I1129 04:56:38.948615 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4bba1b4-2512-4ced-9c5f-43567fd3535b","Type":"ContainerStarted","Data":"241e64d8735438b8ad40484e5da44c33346145df707ef061e3aece3afe572eae"} Nov 29 04:56:39 crc kubenswrapper[4799]: I1129 04:56:39.965466 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4bba1b4-2512-4ced-9c5f-43567fd3535b","Type":"ContainerStarted","Data":"23dbe5855ec057555e17141e858a2e842ddac872d0d628d2dd918666fddfd88c"} Nov 29 04:56:40 crc kubenswrapper[4799]: I1129 04:56:40.755194 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 29 04:56:46 crc kubenswrapper[4799]: I1129 04:56:46.049609 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-5g6rp" event={"ID":"6f53c1df-8c43-4a5a-94c7-7d48fedf34ac","Type":"ContainerStarted","Data":"d2057150feb155f1980728a394f4131b9c9ebd3a0e746e7ccd911fd293524a5c"} Nov 29 04:56:46 crc kubenswrapper[4799]: I1129 04:56:46.063379 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4bba1b4-2512-4ced-9c5f-43567fd3535b","Type":"ContainerStarted","Data":"9d7ecf0c973621d4df309bd7dd169f77e8153ea0d7092387fa53bd47782730bb"} Nov 29 04:56:46 crc kubenswrapper[4799]: I1129 04:56:46.068320 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-5g6rp" podStartSLOduration=2.9991409669999998 podStartE2EDuration="12.068301417s" podCreationTimestamp="2025-11-29 04:56:34 +0000 UTC" firstStartedPulling="2025-11-29 04:56:35.50096271 +0000 UTC m=+1071.143893110" lastFinishedPulling="2025-11-29 04:56:44.57012316 +0000 UTC m=+1080.213053560" observedRunningTime="2025-11-29 04:56:46.064571717 +0000 UTC m=+1081.707502117" watchObservedRunningTime="2025-11-29 04:56:46.068301417 +0000 UTC m=+1081.711231817" Nov 29 04:56:47 crc kubenswrapper[4799]: I1129 04:56:47.080898 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c4bba1b4-2512-4ced-9c5f-43567fd3535b" containerName="ceilometer-central-agent" containerID="cri-o://241e64d8735438b8ad40484e5da44c33346145df707ef061e3aece3afe572eae" gracePeriod=30 Nov 29 04:56:47 crc kubenswrapper[4799]: I1129 04:56:47.082261 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c4bba1b4-2512-4ced-9c5f-43567fd3535b" containerName="proxy-httpd" containerID="cri-o://42330f6d5ce0ff575f6101d3285e729fcde773cc10e0a10c4a4ce30009974d36" gracePeriod=30 Nov 29 04:56:47 crc kubenswrapper[4799]: I1129 04:56:47.082255 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4bba1b4-2512-4ced-9c5f-43567fd3535b","Type":"ContainerStarted","Data":"42330f6d5ce0ff575f6101d3285e729fcde773cc10e0a10c4a4ce30009974d36"} Nov 29 04:56:47 crc kubenswrapper[4799]: I1129 04:56:47.082339 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c4bba1b4-2512-4ced-9c5f-43567fd3535b" containerName="sg-core" containerID="cri-o://9d7ecf0c973621d4df309bd7dd169f77e8153ea0d7092387fa53bd47782730bb" gracePeriod=30 Nov 29 04:56:47 crc kubenswrapper[4799]: I1129 04:56:47.082446 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c4bba1b4-2512-4ced-9c5f-43567fd3535b" containerName="ceilometer-notification-agent" containerID="cri-o://23dbe5855ec057555e17141e858a2e842ddac872d0d628d2dd918666fddfd88c" gracePeriod=30 Nov 29 04:56:47 crc kubenswrapper[4799]: I1129 04:56:47.082612 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 29 04:56:47 crc kubenswrapper[4799]: I1129 04:56:47.109701 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.3211472459999998 podStartE2EDuration="12.109671689s" podCreationTimestamp="2025-11-29 04:56:35 +0000 UTC" firstStartedPulling="2025-11-29 04:56:36.961703444 +0000 UTC m=+1072.604633844" lastFinishedPulling="2025-11-29 04:56:46.750227887 +0000 UTC m=+1082.393158287" observedRunningTime="2025-11-29 04:56:47.10355944 +0000 UTC m=+1082.746489840" watchObservedRunningTime="2025-11-29 04:56:47.109671689 +0000 UTC m=+1082.752602089" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.061022 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.093599 4799 generic.go:334] "Generic (PLEG): container finished" podID="c4bba1b4-2512-4ced-9c5f-43567fd3535b" containerID="42330f6d5ce0ff575f6101d3285e729fcde773cc10e0a10c4a4ce30009974d36" exitCode=0 Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.093644 4799 generic.go:334] "Generic (PLEG): container finished" podID="c4bba1b4-2512-4ced-9c5f-43567fd3535b" containerID="9d7ecf0c973621d4df309bd7dd169f77e8153ea0d7092387fa53bd47782730bb" exitCode=2 Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.093654 4799 generic.go:334] "Generic (PLEG): container finished" podID="c4bba1b4-2512-4ced-9c5f-43567fd3535b" containerID="23dbe5855ec057555e17141e858a2e842ddac872d0d628d2dd918666fddfd88c" exitCode=0 Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.093664 4799 generic.go:334] "Generic (PLEG): container finished" podID="c4bba1b4-2512-4ced-9c5f-43567fd3535b" containerID="241e64d8735438b8ad40484e5da44c33346145df707ef061e3aece3afe572eae" exitCode=0 Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.093692 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4bba1b4-2512-4ced-9c5f-43567fd3535b","Type":"ContainerDied","Data":"42330f6d5ce0ff575f6101d3285e729fcde773cc10e0a10c4a4ce30009974d36"} Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.093728 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.093771 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4bba1b4-2512-4ced-9c5f-43567fd3535b","Type":"ContainerDied","Data":"9d7ecf0c973621d4df309bd7dd169f77e8153ea0d7092387fa53bd47782730bb"} Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.093813 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4bba1b4-2512-4ced-9c5f-43567fd3535b","Type":"ContainerDied","Data":"23dbe5855ec057555e17141e858a2e842ddac872d0d628d2dd918666fddfd88c"} Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.093833 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4bba1b4-2512-4ced-9c5f-43567fd3535b","Type":"ContainerDied","Data":"241e64d8735438b8ad40484e5da44c33346145df707ef061e3aece3afe572eae"} Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.093850 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4bba1b4-2512-4ced-9c5f-43567fd3535b","Type":"ContainerDied","Data":"b61227d6a655050ee38dc18d24bb88c8f84832b215c9dca2b78b4391c4e52ed8"} Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.093877 4799 scope.go:117] "RemoveContainer" containerID="42330f6d5ce0ff575f6101d3285e729fcde773cc10e0a10c4a4ce30009974d36" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.120252 4799 scope.go:117] "RemoveContainer" containerID="9d7ecf0c973621d4df309bd7dd169f77e8153ea0d7092387fa53bd47782730bb" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.140074 4799 scope.go:117] "RemoveContainer" containerID="23dbe5855ec057555e17141e858a2e842ddac872d0d628d2dd918666fddfd88c" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.156625 4799 scope.go:117] "RemoveContainer" containerID="241e64d8735438b8ad40484e5da44c33346145df707ef061e3aece3afe572eae" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.176556 4799 scope.go:117] "RemoveContainer" containerID="42330f6d5ce0ff575f6101d3285e729fcde773cc10e0a10c4a4ce30009974d36" Nov 29 04:56:48 crc kubenswrapper[4799]: E1129 04:56:48.177075 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42330f6d5ce0ff575f6101d3285e729fcde773cc10e0a10c4a4ce30009974d36\": container with ID starting with 42330f6d5ce0ff575f6101d3285e729fcde773cc10e0a10c4a4ce30009974d36 not found: ID does not exist" containerID="42330f6d5ce0ff575f6101d3285e729fcde773cc10e0a10c4a4ce30009974d36" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.177134 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42330f6d5ce0ff575f6101d3285e729fcde773cc10e0a10c4a4ce30009974d36"} err="failed to get container status \"42330f6d5ce0ff575f6101d3285e729fcde773cc10e0a10c4a4ce30009974d36\": rpc error: code = NotFound desc = could not find container \"42330f6d5ce0ff575f6101d3285e729fcde773cc10e0a10c4a4ce30009974d36\": container with ID starting with 42330f6d5ce0ff575f6101d3285e729fcde773cc10e0a10c4a4ce30009974d36 not found: ID does not exist" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.177166 4799 scope.go:117] "RemoveContainer" containerID="9d7ecf0c973621d4df309bd7dd169f77e8153ea0d7092387fa53bd47782730bb" Nov 29 04:56:48 crc kubenswrapper[4799]: E1129 04:56:48.177453 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d7ecf0c973621d4df309bd7dd169f77e8153ea0d7092387fa53bd47782730bb\": container with ID starting with 9d7ecf0c973621d4df309bd7dd169f77e8153ea0d7092387fa53bd47782730bb not found: ID does not exist" containerID="9d7ecf0c973621d4df309bd7dd169f77e8153ea0d7092387fa53bd47782730bb" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.177486 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d7ecf0c973621d4df309bd7dd169f77e8153ea0d7092387fa53bd47782730bb"} err="failed to get container status \"9d7ecf0c973621d4df309bd7dd169f77e8153ea0d7092387fa53bd47782730bb\": rpc error: code = NotFound desc = could not find container \"9d7ecf0c973621d4df309bd7dd169f77e8153ea0d7092387fa53bd47782730bb\": container with ID starting with 9d7ecf0c973621d4df309bd7dd169f77e8153ea0d7092387fa53bd47782730bb not found: ID does not exist" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.177508 4799 scope.go:117] "RemoveContainer" containerID="23dbe5855ec057555e17141e858a2e842ddac872d0d628d2dd918666fddfd88c" Nov 29 04:56:48 crc kubenswrapper[4799]: E1129 04:56:48.177747 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23dbe5855ec057555e17141e858a2e842ddac872d0d628d2dd918666fddfd88c\": container with ID starting with 23dbe5855ec057555e17141e858a2e842ddac872d0d628d2dd918666fddfd88c not found: ID does not exist" containerID="23dbe5855ec057555e17141e858a2e842ddac872d0d628d2dd918666fddfd88c" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.177808 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23dbe5855ec057555e17141e858a2e842ddac872d0d628d2dd918666fddfd88c"} err="failed to get container status \"23dbe5855ec057555e17141e858a2e842ddac872d0d628d2dd918666fddfd88c\": rpc error: code = NotFound desc = could not find container \"23dbe5855ec057555e17141e858a2e842ddac872d0d628d2dd918666fddfd88c\": container with ID starting with 23dbe5855ec057555e17141e858a2e842ddac872d0d628d2dd918666fddfd88c not found: ID does not exist" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.177831 4799 scope.go:117] "RemoveContainer" containerID="241e64d8735438b8ad40484e5da44c33346145df707ef061e3aece3afe572eae" Nov 29 04:56:48 crc kubenswrapper[4799]: E1129 04:56:48.178120 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"241e64d8735438b8ad40484e5da44c33346145df707ef061e3aece3afe572eae\": container with ID starting with 241e64d8735438b8ad40484e5da44c33346145df707ef061e3aece3afe572eae not found: ID does not exist" containerID="241e64d8735438b8ad40484e5da44c33346145df707ef061e3aece3afe572eae" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.178149 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"241e64d8735438b8ad40484e5da44c33346145df707ef061e3aece3afe572eae"} err="failed to get container status \"241e64d8735438b8ad40484e5da44c33346145df707ef061e3aece3afe572eae\": rpc error: code = NotFound desc = could not find container \"241e64d8735438b8ad40484e5da44c33346145df707ef061e3aece3afe572eae\": container with ID starting with 241e64d8735438b8ad40484e5da44c33346145df707ef061e3aece3afe572eae not found: ID does not exist" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.178161 4799 scope.go:117] "RemoveContainer" containerID="42330f6d5ce0ff575f6101d3285e729fcde773cc10e0a10c4a4ce30009974d36" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.178428 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42330f6d5ce0ff575f6101d3285e729fcde773cc10e0a10c4a4ce30009974d36"} err="failed to get container status \"42330f6d5ce0ff575f6101d3285e729fcde773cc10e0a10c4a4ce30009974d36\": rpc error: code = NotFound desc = could not find container \"42330f6d5ce0ff575f6101d3285e729fcde773cc10e0a10c4a4ce30009974d36\": container with ID starting with 42330f6d5ce0ff575f6101d3285e729fcde773cc10e0a10c4a4ce30009974d36 not found: ID does not exist" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.178459 4799 scope.go:117] "RemoveContainer" containerID="9d7ecf0c973621d4df309bd7dd169f77e8153ea0d7092387fa53bd47782730bb" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.178724 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d7ecf0c973621d4df309bd7dd169f77e8153ea0d7092387fa53bd47782730bb"} err="failed to get container status \"9d7ecf0c973621d4df309bd7dd169f77e8153ea0d7092387fa53bd47782730bb\": rpc error: code = NotFound desc = could not find container \"9d7ecf0c973621d4df309bd7dd169f77e8153ea0d7092387fa53bd47782730bb\": container with ID starting with 9d7ecf0c973621d4df309bd7dd169f77e8153ea0d7092387fa53bd47782730bb not found: ID does not exist" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.178754 4799 scope.go:117] "RemoveContainer" containerID="23dbe5855ec057555e17141e858a2e842ddac872d0d628d2dd918666fddfd88c" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.179106 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23dbe5855ec057555e17141e858a2e842ddac872d0d628d2dd918666fddfd88c"} err="failed to get container status \"23dbe5855ec057555e17141e858a2e842ddac872d0d628d2dd918666fddfd88c\": rpc error: code = NotFound desc = could not find container \"23dbe5855ec057555e17141e858a2e842ddac872d0d628d2dd918666fddfd88c\": container with ID starting with 23dbe5855ec057555e17141e858a2e842ddac872d0d628d2dd918666fddfd88c not found: ID does not exist" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.179126 4799 scope.go:117] "RemoveContainer" containerID="241e64d8735438b8ad40484e5da44c33346145df707ef061e3aece3afe572eae" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.179520 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"241e64d8735438b8ad40484e5da44c33346145df707ef061e3aece3afe572eae"} err="failed to get container status \"241e64d8735438b8ad40484e5da44c33346145df707ef061e3aece3afe572eae\": rpc error: code = NotFound desc = could not find container \"241e64d8735438b8ad40484e5da44c33346145df707ef061e3aece3afe572eae\": container with ID starting with 241e64d8735438b8ad40484e5da44c33346145df707ef061e3aece3afe572eae not found: ID does not exist" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.179587 4799 scope.go:117] "RemoveContainer" containerID="42330f6d5ce0ff575f6101d3285e729fcde773cc10e0a10c4a4ce30009974d36" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.179923 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42330f6d5ce0ff575f6101d3285e729fcde773cc10e0a10c4a4ce30009974d36"} err="failed to get container status \"42330f6d5ce0ff575f6101d3285e729fcde773cc10e0a10c4a4ce30009974d36\": rpc error: code = NotFound desc = could not find container \"42330f6d5ce0ff575f6101d3285e729fcde773cc10e0a10c4a4ce30009974d36\": container with ID starting with 42330f6d5ce0ff575f6101d3285e729fcde773cc10e0a10c4a4ce30009974d36 not found: ID does not exist" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.180013 4799 scope.go:117] "RemoveContainer" containerID="9d7ecf0c973621d4df309bd7dd169f77e8153ea0d7092387fa53bd47782730bb" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.180325 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d7ecf0c973621d4df309bd7dd169f77e8153ea0d7092387fa53bd47782730bb"} err="failed to get container status \"9d7ecf0c973621d4df309bd7dd169f77e8153ea0d7092387fa53bd47782730bb\": rpc error: code = NotFound desc = could not find container \"9d7ecf0c973621d4df309bd7dd169f77e8153ea0d7092387fa53bd47782730bb\": container with ID starting with 9d7ecf0c973621d4df309bd7dd169f77e8153ea0d7092387fa53bd47782730bb not found: ID does not exist" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.180351 4799 scope.go:117] "RemoveContainer" containerID="23dbe5855ec057555e17141e858a2e842ddac872d0d628d2dd918666fddfd88c" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.180605 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23dbe5855ec057555e17141e858a2e842ddac872d0d628d2dd918666fddfd88c"} err="failed to get container status \"23dbe5855ec057555e17141e858a2e842ddac872d0d628d2dd918666fddfd88c\": rpc error: code = NotFound desc = could not find container \"23dbe5855ec057555e17141e858a2e842ddac872d0d628d2dd918666fddfd88c\": container with ID starting with 23dbe5855ec057555e17141e858a2e842ddac872d0d628d2dd918666fddfd88c not found: ID does not exist" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.180630 4799 scope.go:117] "RemoveContainer" containerID="241e64d8735438b8ad40484e5da44c33346145df707ef061e3aece3afe572eae" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.180932 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"241e64d8735438b8ad40484e5da44c33346145df707ef061e3aece3afe572eae"} err="failed to get container status \"241e64d8735438b8ad40484e5da44c33346145df707ef061e3aece3afe572eae\": rpc error: code = NotFound desc = could not find container \"241e64d8735438b8ad40484e5da44c33346145df707ef061e3aece3afe572eae\": container with ID starting with 241e64d8735438b8ad40484e5da44c33346145df707ef061e3aece3afe572eae not found: ID does not exist" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.180970 4799 scope.go:117] "RemoveContainer" containerID="42330f6d5ce0ff575f6101d3285e729fcde773cc10e0a10c4a4ce30009974d36" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.181395 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42330f6d5ce0ff575f6101d3285e729fcde773cc10e0a10c4a4ce30009974d36"} err="failed to get container status \"42330f6d5ce0ff575f6101d3285e729fcde773cc10e0a10c4a4ce30009974d36\": rpc error: code = NotFound desc = could not find container \"42330f6d5ce0ff575f6101d3285e729fcde773cc10e0a10c4a4ce30009974d36\": container with ID starting with 42330f6d5ce0ff575f6101d3285e729fcde773cc10e0a10c4a4ce30009974d36 not found: ID does not exist" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.181422 4799 scope.go:117] "RemoveContainer" containerID="9d7ecf0c973621d4df309bd7dd169f77e8153ea0d7092387fa53bd47782730bb" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.181671 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d7ecf0c973621d4df309bd7dd169f77e8153ea0d7092387fa53bd47782730bb"} err="failed to get container status \"9d7ecf0c973621d4df309bd7dd169f77e8153ea0d7092387fa53bd47782730bb\": rpc error: code = NotFound desc = could not find container \"9d7ecf0c973621d4df309bd7dd169f77e8153ea0d7092387fa53bd47782730bb\": container with ID starting with 9d7ecf0c973621d4df309bd7dd169f77e8153ea0d7092387fa53bd47782730bb not found: ID does not exist" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.181702 4799 scope.go:117] "RemoveContainer" containerID="23dbe5855ec057555e17141e858a2e842ddac872d0d628d2dd918666fddfd88c" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.182190 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23dbe5855ec057555e17141e858a2e842ddac872d0d628d2dd918666fddfd88c"} err="failed to get container status \"23dbe5855ec057555e17141e858a2e842ddac872d0d628d2dd918666fddfd88c\": rpc error: code = NotFound desc = could not find container \"23dbe5855ec057555e17141e858a2e842ddac872d0d628d2dd918666fddfd88c\": container with ID starting with 23dbe5855ec057555e17141e858a2e842ddac872d0d628d2dd918666fddfd88c not found: ID does not exist" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.182315 4799 scope.go:117] "RemoveContainer" containerID="241e64d8735438b8ad40484e5da44c33346145df707ef061e3aece3afe572eae" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.182730 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"241e64d8735438b8ad40484e5da44c33346145df707ef061e3aece3afe572eae"} err="failed to get container status \"241e64d8735438b8ad40484e5da44c33346145df707ef061e3aece3afe572eae\": rpc error: code = NotFound desc = could not find container \"241e64d8735438b8ad40484e5da44c33346145df707ef061e3aece3afe572eae\": container with ID starting with 241e64d8735438b8ad40484e5da44c33346145df707ef061e3aece3afe572eae not found: ID does not exist" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.190751 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4bba1b4-2512-4ced-9c5f-43567fd3535b-log-httpd\") pod \"c4bba1b4-2512-4ced-9c5f-43567fd3535b\" (UID: \"c4bba1b4-2512-4ced-9c5f-43567fd3535b\") " Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.190912 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4bba1b4-2512-4ced-9c5f-43567fd3535b-combined-ca-bundle\") pod \"c4bba1b4-2512-4ced-9c5f-43567fd3535b\" (UID: \"c4bba1b4-2512-4ced-9c5f-43567fd3535b\") " Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.190950 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c4bba1b4-2512-4ced-9c5f-43567fd3535b-sg-core-conf-yaml\") pod \"c4bba1b4-2512-4ced-9c5f-43567fd3535b\" (UID: \"c4bba1b4-2512-4ced-9c5f-43567fd3535b\") " Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.190977 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-62v52\" (UniqueName: \"kubernetes.io/projected/c4bba1b4-2512-4ced-9c5f-43567fd3535b-kube-api-access-62v52\") pod \"c4bba1b4-2512-4ced-9c5f-43567fd3535b\" (UID: \"c4bba1b4-2512-4ced-9c5f-43567fd3535b\") " Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.191019 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4bba1b4-2512-4ced-9c5f-43567fd3535b-run-httpd\") pod \"c4bba1b4-2512-4ced-9c5f-43567fd3535b\" (UID: \"c4bba1b4-2512-4ced-9c5f-43567fd3535b\") " Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.191498 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4bba1b4-2512-4ced-9c5f-43567fd3535b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c4bba1b4-2512-4ced-9c5f-43567fd3535b" (UID: "c4bba1b4-2512-4ced-9c5f-43567fd3535b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.191808 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4bba1b4-2512-4ced-9c5f-43567fd3535b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c4bba1b4-2512-4ced-9c5f-43567fd3535b" (UID: "c4bba1b4-2512-4ced-9c5f-43567fd3535b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.192146 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4bba1b4-2512-4ced-9c5f-43567fd3535b-config-data\") pod \"c4bba1b4-2512-4ced-9c5f-43567fd3535b\" (UID: \"c4bba1b4-2512-4ced-9c5f-43567fd3535b\") " Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.192562 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4bba1b4-2512-4ced-9c5f-43567fd3535b-scripts\") pod \"c4bba1b4-2512-4ced-9c5f-43567fd3535b\" (UID: \"c4bba1b4-2512-4ced-9c5f-43567fd3535b\") " Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.193410 4799 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4bba1b4-2512-4ced-9c5f-43567fd3535b-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.193445 4799 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4bba1b4-2512-4ced-9c5f-43567fd3535b-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.196829 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4bba1b4-2512-4ced-9c5f-43567fd3535b-scripts" (OuterVolumeSpecName: "scripts") pod "c4bba1b4-2512-4ced-9c5f-43567fd3535b" (UID: "c4bba1b4-2512-4ced-9c5f-43567fd3535b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.197370 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4bba1b4-2512-4ced-9c5f-43567fd3535b-kube-api-access-62v52" (OuterVolumeSpecName: "kube-api-access-62v52") pod "c4bba1b4-2512-4ced-9c5f-43567fd3535b" (UID: "c4bba1b4-2512-4ced-9c5f-43567fd3535b"). InnerVolumeSpecName "kube-api-access-62v52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.216738 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4bba1b4-2512-4ced-9c5f-43567fd3535b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c4bba1b4-2512-4ced-9c5f-43567fd3535b" (UID: "c4bba1b4-2512-4ced-9c5f-43567fd3535b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.257159 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4bba1b4-2512-4ced-9c5f-43567fd3535b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c4bba1b4-2512-4ced-9c5f-43567fd3535b" (UID: "c4bba1b4-2512-4ced-9c5f-43567fd3535b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.288063 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4bba1b4-2512-4ced-9c5f-43567fd3535b-config-data" (OuterVolumeSpecName: "config-data") pod "c4bba1b4-2512-4ced-9c5f-43567fd3535b" (UID: "c4bba1b4-2512-4ced-9c5f-43567fd3535b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.295657 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4bba1b4-2512-4ced-9c5f-43567fd3535b-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.295686 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4bba1b4-2512-4ced-9c5f-43567fd3535b-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.295702 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4bba1b4-2512-4ced-9c5f-43567fd3535b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.295716 4799 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c4bba1b4-2512-4ced-9c5f-43567fd3535b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.295734 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-62v52\" (UniqueName: \"kubernetes.io/projected/c4bba1b4-2512-4ced-9c5f-43567fd3535b-kube-api-access-62v52\") on node \"crc\" DevicePath \"\"" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.435370 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.448532 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.479215 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 29 04:56:48 crc kubenswrapper[4799]: E1129 04:56:48.479707 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4bba1b4-2512-4ced-9c5f-43567fd3535b" containerName="sg-core" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.479730 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4bba1b4-2512-4ced-9c5f-43567fd3535b" containerName="sg-core" Nov 29 04:56:48 crc kubenswrapper[4799]: E1129 04:56:48.479748 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4bba1b4-2512-4ced-9c5f-43567fd3535b" containerName="ceilometer-central-agent" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.479755 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4bba1b4-2512-4ced-9c5f-43567fd3535b" containerName="ceilometer-central-agent" Nov 29 04:56:48 crc kubenswrapper[4799]: E1129 04:56:48.479769 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4bba1b4-2512-4ced-9c5f-43567fd3535b" containerName="proxy-httpd" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.479777 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4bba1b4-2512-4ced-9c5f-43567fd3535b" containerName="proxy-httpd" Nov 29 04:56:48 crc kubenswrapper[4799]: E1129 04:56:48.479811 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4bba1b4-2512-4ced-9c5f-43567fd3535b" containerName="ceilometer-notification-agent" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.479818 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4bba1b4-2512-4ced-9c5f-43567fd3535b" containerName="ceilometer-notification-agent" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.479974 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4bba1b4-2512-4ced-9c5f-43567fd3535b" containerName="sg-core" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.479994 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4bba1b4-2512-4ced-9c5f-43567fd3535b" containerName="ceilometer-central-agent" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.480004 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4bba1b4-2512-4ced-9c5f-43567fd3535b" containerName="ceilometer-notification-agent" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.480018 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4bba1b4-2512-4ced-9c5f-43567fd3535b" containerName="proxy-httpd" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.481852 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.486834 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.486837 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.505154 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.599669 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d4741f22-658c-40af-bba0-b36c491be636-run-httpd\") pod \"ceilometer-0\" (UID: \"d4741f22-658c-40af-bba0-b36c491be636\") " pod="openstack/ceilometer-0" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.599739 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qht5x\" (UniqueName: \"kubernetes.io/projected/d4741f22-658c-40af-bba0-b36c491be636-kube-api-access-qht5x\") pod \"ceilometer-0\" (UID: \"d4741f22-658c-40af-bba0-b36c491be636\") " pod="openstack/ceilometer-0" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.599807 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4741f22-658c-40af-bba0-b36c491be636-scripts\") pod \"ceilometer-0\" (UID: \"d4741f22-658c-40af-bba0-b36c491be636\") " pod="openstack/ceilometer-0" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.599855 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4741f22-658c-40af-bba0-b36c491be636-config-data\") pod \"ceilometer-0\" (UID: \"d4741f22-658c-40af-bba0-b36c491be636\") " pod="openstack/ceilometer-0" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.599956 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4741f22-658c-40af-bba0-b36c491be636-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d4741f22-658c-40af-bba0-b36c491be636\") " pod="openstack/ceilometer-0" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.599973 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d4741f22-658c-40af-bba0-b36c491be636-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d4741f22-658c-40af-bba0-b36c491be636\") " pod="openstack/ceilometer-0" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.599997 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d4741f22-658c-40af-bba0-b36c491be636-log-httpd\") pod \"ceilometer-0\" (UID: \"d4741f22-658c-40af-bba0-b36c491be636\") " pod="openstack/ceilometer-0" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.668475 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4bba1b4-2512-4ced-9c5f-43567fd3535b" path="/var/lib/kubelet/pods/c4bba1b4-2512-4ced-9c5f-43567fd3535b/volumes" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.702114 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4741f22-658c-40af-bba0-b36c491be636-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d4741f22-658c-40af-bba0-b36c491be636\") " pod="openstack/ceilometer-0" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.702158 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d4741f22-658c-40af-bba0-b36c491be636-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d4741f22-658c-40af-bba0-b36c491be636\") " pod="openstack/ceilometer-0" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.702195 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d4741f22-658c-40af-bba0-b36c491be636-log-httpd\") pod \"ceilometer-0\" (UID: \"d4741f22-658c-40af-bba0-b36c491be636\") " pod="openstack/ceilometer-0" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.702239 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d4741f22-658c-40af-bba0-b36c491be636-run-httpd\") pod \"ceilometer-0\" (UID: \"d4741f22-658c-40af-bba0-b36c491be636\") " pod="openstack/ceilometer-0" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.702273 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qht5x\" (UniqueName: \"kubernetes.io/projected/d4741f22-658c-40af-bba0-b36c491be636-kube-api-access-qht5x\") pod \"ceilometer-0\" (UID: \"d4741f22-658c-40af-bba0-b36c491be636\") " pod="openstack/ceilometer-0" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.702299 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4741f22-658c-40af-bba0-b36c491be636-scripts\") pod \"ceilometer-0\" (UID: \"d4741f22-658c-40af-bba0-b36c491be636\") " pod="openstack/ceilometer-0" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.702333 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4741f22-658c-40af-bba0-b36c491be636-config-data\") pod \"ceilometer-0\" (UID: \"d4741f22-658c-40af-bba0-b36c491be636\") " pod="openstack/ceilometer-0" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.702935 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d4741f22-658c-40af-bba0-b36c491be636-run-httpd\") pod \"ceilometer-0\" (UID: \"d4741f22-658c-40af-bba0-b36c491be636\") " pod="openstack/ceilometer-0" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.703078 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d4741f22-658c-40af-bba0-b36c491be636-log-httpd\") pod \"ceilometer-0\" (UID: \"d4741f22-658c-40af-bba0-b36c491be636\") " pod="openstack/ceilometer-0" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.708313 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d4741f22-658c-40af-bba0-b36c491be636-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d4741f22-658c-40af-bba0-b36c491be636\") " pod="openstack/ceilometer-0" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.708820 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4741f22-658c-40af-bba0-b36c491be636-config-data\") pod \"ceilometer-0\" (UID: \"d4741f22-658c-40af-bba0-b36c491be636\") " pod="openstack/ceilometer-0" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.708846 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4741f22-658c-40af-bba0-b36c491be636-scripts\") pod \"ceilometer-0\" (UID: \"d4741f22-658c-40af-bba0-b36c491be636\") " pod="openstack/ceilometer-0" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.709726 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4741f22-658c-40af-bba0-b36c491be636-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d4741f22-658c-40af-bba0-b36c491be636\") " pod="openstack/ceilometer-0" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.722358 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qht5x\" (UniqueName: \"kubernetes.io/projected/d4741f22-658c-40af-bba0-b36c491be636-kube-api-access-qht5x\") pod \"ceilometer-0\" (UID: \"d4741f22-658c-40af-bba0-b36c491be636\") " pod="openstack/ceilometer-0" Nov 29 04:56:48 crc kubenswrapper[4799]: I1129 04:56:48.805987 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 04:56:49 crc kubenswrapper[4799]: I1129 04:56:49.257075 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 29 04:56:50 crc kubenswrapper[4799]: I1129 04:56:50.119446 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d4741f22-658c-40af-bba0-b36c491be636","Type":"ContainerStarted","Data":"cb2cd9b60acdc3381d394780aa0fc5cf6c8d8b1a94029038393ac04292fbfb98"} Nov 29 04:56:50 crc kubenswrapper[4799]: I1129 04:56:50.119851 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d4741f22-658c-40af-bba0-b36c491be636","Type":"ContainerStarted","Data":"30734b3a95d85ea21cbe9e0d1f2c140cb31d17f67cf4b2cf8e610d4be32d771e"} Nov 29 04:56:51 crc kubenswrapper[4799]: I1129 04:56:51.131608 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d4741f22-658c-40af-bba0-b36c491be636","Type":"ContainerStarted","Data":"ff92573600da0847b820dabc90979d71078d06a62fe0d43ca46e4ca0b36f4328"} Nov 29 04:56:52 crc kubenswrapper[4799]: I1129 04:56:52.143750 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d4741f22-658c-40af-bba0-b36c491be636","Type":"ContainerStarted","Data":"bc58c8ff01459001f39e28a7979a9c4bfd0e961f421845b23dd2cff71163ceb3"} Nov 29 04:56:54 crc kubenswrapper[4799]: I1129 04:56:54.199781 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d4741f22-658c-40af-bba0-b36c491be636","Type":"ContainerStarted","Data":"d67f65d1c9b133f52aae7c40e3e167681a53206dbdf3f6cf9575afc01f64f4b6"} Nov 29 04:56:54 crc kubenswrapper[4799]: I1129 04:56:54.201082 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 29 04:56:54 crc kubenswrapper[4799]: I1129 04:56:54.231916 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.593275353 podStartE2EDuration="6.231893521s" podCreationTimestamp="2025-11-29 04:56:48 +0000 UTC" firstStartedPulling="2025-11-29 04:56:49.273275183 +0000 UTC m=+1084.916205573" lastFinishedPulling="2025-11-29 04:56:52.911893321 +0000 UTC m=+1088.554823741" observedRunningTime="2025-11-29 04:56:54.22728972 +0000 UTC m=+1089.870220120" watchObservedRunningTime="2025-11-29 04:56:54.231893521 +0000 UTC m=+1089.874823921" Nov 29 04:56:59 crc kubenswrapper[4799]: I1129 04:56:59.259944 4799 generic.go:334] "Generic (PLEG): container finished" podID="6f53c1df-8c43-4a5a-94c7-7d48fedf34ac" containerID="d2057150feb155f1980728a394f4131b9c9ebd3a0e746e7ccd911fd293524a5c" exitCode=0 Nov 29 04:56:59 crc kubenswrapper[4799]: I1129 04:56:59.260117 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-5g6rp" event={"ID":"6f53c1df-8c43-4a5a-94c7-7d48fedf34ac","Type":"ContainerDied","Data":"d2057150feb155f1980728a394f4131b9c9ebd3a0e746e7ccd911fd293524a5c"} Nov 29 04:57:00 crc kubenswrapper[4799]: I1129 04:57:00.686918 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-5g6rp" Nov 29 04:57:00 crc kubenswrapper[4799]: I1129 04:57:00.772406 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5l2ql\" (UniqueName: \"kubernetes.io/projected/6f53c1df-8c43-4a5a-94c7-7d48fedf34ac-kube-api-access-5l2ql\") pod \"6f53c1df-8c43-4a5a-94c7-7d48fedf34ac\" (UID: \"6f53c1df-8c43-4a5a-94c7-7d48fedf34ac\") " Nov 29 04:57:00 crc kubenswrapper[4799]: I1129 04:57:00.772598 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f53c1df-8c43-4a5a-94c7-7d48fedf34ac-combined-ca-bundle\") pod \"6f53c1df-8c43-4a5a-94c7-7d48fedf34ac\" (UID: \"6f53c1df-8c43-4a5a-94c7-7d48fedf34ac\") " Nov 29 04:57:00 crc kubenswrapper[4799]: I1129 04:57:00.772703 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f53c1df-8c43-4a5a-94c7-7d48fedf34ac-scripts\") pod \"6f53c1df-8c43-4a5a-94c7-7d48fedf34ac\" (UID: \"6f53c1df-8c43-4a5a-94c7-7d48fedf34ac\") " Nov 29 04:57:00 crc kubenswrapper[4799]: I1129 04:57:00.772781 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f53c1df-8c43-4a5a-94c7-7d48fedf34ac-config-data\") pod \"6f53c1df-8c43-4a5a-94c7-7d48fedf34ac\" (UID: \"6f53c1df-8c43-4a5a-94c7-7d48fedf34ac\") " Nov 29 04:57:00 crc kubenswrapper[4799]: I1129 04:57:00.783520 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f53c1df-8c43-4a5a-94c7-7d48fedf34ac-scripts" (OuterVolumeSpecName: "scripts") pod "6f53c1df-8c43-4a5a-94c7-7d48fedf34ac" (UID: "6f53c1df-8c43-4a5a-94c7-7d48fedf34ac"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:57:00 crc kubenswrapper[4799]: I1129 04:57:00.783658 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f53c1df-8c43-4a5a-94c7-7d48fedf34ac-kube-api-access-5l2ql" (OuterVolumeSpecName: "kube-api-access-5l2ql") pod "6f53c1df-8c43-4a5a-94c7-7d48fedf34ac" (UID: "6f53c1df-8c43-4a5a-94c7-7d48fedf34ac"). InnerVolumeSpecName "kube-api-access-5l2ql". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:57:00 crc kubenswrapper[4799]: I1129 04:57:00.814975 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f53c1df-8c43-4a5a-94c7-7d48fedf34ac-config-data" (OuterVolumeSpecName: "config-data") pod "6f53c1df-8c43-4a5a-94c7-7d48fedf34ac" (UID: "6f53c1df-8c43-4a5a-94c7-7d48fedf34ac"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:57:00 crc kubenswrapper[4799]: I1129 04:57:00.825916 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f53c1df-8c43-4a5a-94c7-7d48fedf34ac-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6f53c1df-8c43-4a5a-94c7-7d48fedf34ac" (UID: "6f53c1df-8c43-4a5a-94c7-7d48fedf34ac"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:57:00 crc kubenswrapper[4799]: I1129 04:57:00.875155 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5l2ql\" (UniqueName: \"kubernetes.io/projected/6f53c1df-8c43-4a5a-94c7-7d48fedf34ac-kube-api-access-5l2ql\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:00 crc kubenswrapper[4799]: I1129 04:57:00.875184 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f53c1df-8c43-4a5a-94c7-7d48fedf34ac-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:00 crc kubenswrapper[4799]: I1129 04:57:00.875195 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f53c1df-8c43-4a5a-94c7-7d48fedf34ac-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:00 crc kubenswrapper[4799]: I1129 04:57:00.875204 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f53c1df-8c43-4a5a-94c7-7d48fedf34ac-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:01 crc kubenswrapper[4799]: I1129 04:57:01.286153 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-5g6rp" event={"ID":"6f53c1df-8c43-4a5a-94c7-7d48fedf34ac","Type":"ContainerDied","Data":"330d4eca3cb2178385d83fbf34af6bb57705209a177a56249e53527480d404da"} Nov 29 04:57:01 crc kubenswrapper[4799]: I1129 04:57:01.286553 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="330d4eca3cb2178385d83fbf34af6bb57705209a177a56249e53527480d404da" Nov 29 04:57:01 crc kubenswrapper[4799]: I1129 04:57:01.286208 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-5g6rp" Nov 29 04:57:01 crc kubenswrapper[4799]: I1129 04:57:01.395527 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 29 04:57:01 crc kubenswrapper[4799]: E1129 04:57:01.396183 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f53c1df-8c43-4a5a-94c7-7d48fedf34ac" containerName="nova-cell0-conductor-db-sync" Nov 29 04:57:01 crc kubenswrapper[4799]: I1129 04:57:01.396205 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f53c1df-8c43-4a5a-94c7-7d48fedf34ac" containerName="nova-cell0-conductor-db-sync" Nov 29 04:57:01 crc kubenswrapper[4799]: I1129 04:57:01.396397 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f53c1df-8c43-4a5a-94c7-7d48fedf34ac" containerName="nova-cell0-conductor-db-sync" Nov 29 04:57:01 crc kubenswrapper[4799]: I1129 04:57:01.397371 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 29 04:57:01 crc kubenswrapper[4799]: I1129 04:57:01.400051 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-5hbsz" Nov 29 04:57:01 crc kubenswrapper[4799]: I1129 04:57:01.413037 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 29 04:57:01 crc kubenswrapper[4799]: I1129 04:57:01.414878 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 29 04:57:01 crc kubenswrapper[4799]: I1129 04:57:01.487715 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6a2c6d8-ac5d-4b14-a907-d624f22d5f1b-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"f6a2c6d8-ac5d-4b14-a907-d624f22d5f1b\") " pod="openstack/nova-cell0-conductor-0" Nov 29 04:57:01 crc kubenswrapper[4799]: I1129 04:57:01.487805 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6a2c6d8-ac5d-4b14-a907-d624f22d5f1b-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"f6a2c6d8-ac5d-4b14-a907-d624f22d5f1b\") " pod="openstack/nova-cell0-conductor-0" Nov 29 04:57:01 crc kubenswrapper[4799]: I1129 04:57:01.488236 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxm2t\" (UniqueName: \"kubernetes.io/projected/f6a2c6d8-ac5d-4b14-a907-d624f22d5f1b-kube-api-access-bxm2t\") pod \"nova-cell0-conductor-0\" (UID: \"f6a2c6d8-ac5d-4b14-a907-d624f22d5f1b\") " pod="openstack/nova-cell0-conductor-0" Nov 29 04:57:01 crc kubenswrapper[4799]: I1129 04:57:01.590601 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxm2t\" (UniqueName: \"kubernetes.io/projected/f6a2c6d8-ac5d-4b14-a907-d624f22d5f1b-kube-api-access-bxm2t\") pod \"nova-cell0-conductor-0\" (UID: \"f6a2c6d8-ac5d-4b14-a907-d624f22d5f1b\") " pod="openstack/nova-cell0-conductor-0" Nov 29 04:57:01 crc kubenswrapper[4799]: I1129 04:57:01.590675 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6a2c6d8-ac5d-4b14-a907-d624f22d5f1b-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"f6a2c6d8-ac5d-4b14-a907-d624f22d5f1b\") " pod="openstack/nova-cell0-conductor-0" Nov 29 04:57:01 crc kubenswrapper[4799]: I1129 04:57:01.590724 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6a2c6d8-ac5d-4b14-a907-d624f22d5f1b-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"f6a2c6d8-ac5d-4b14-a907-d624f22d5f1b\") " pod="openstack/nova-cell0-conductor-0" Nov 29 04:57:01 crc kubenswrapper[4799]: I1129 04:57:01.597160 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6a2c6d8-ac5d-4b14-a907-d624f22d5f1b-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"f6a2c6d8-ac5d-4b14-a907-d624f22d5f1b\") " pod="openstack/nova-cell0-conductor-0" Nov 29 04:57:01 crc kubenswrapper[4799]: I1129 04:57:01.605304 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6a2c6d8-ac5d-4b14-a907-d624f22d5f1b-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"f6a2c6d8-ac5d-4b14-a907-d624f22d5f1b\") " pod="openstack/nova-cell0-conductor-0" Nov 29 04:57:01 crc kubenswrapper[4799]: I1129 04:57:01.615368 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxm2t\" (UniqueName: \"kubernetes.io/projected/f6a2c6d8-ac5d-4b14-a907-d624f22d5f1b-kube-api-access-bxm2t\") pod \"nova-cell0-conductor-0\" (UID: \"f6a2c6d8-ac5d-4b14-a907-d624f22d5f1b\") " pod="openstack/nova-cell0-conductor-0" Nov 29 04:57:01 crc kubenswrapper[4799]: I1129 04:57:01.715896 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 29 04:57:02 crc kubenswrapper[4799]: I1129 04:57:02.230914 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 29 04:57:02 crc kubenswrapper[4799]: I1129 04:57:02.306606 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"f6a2c6d8-ac5d-4b14-a907-d624f22d5f1b","Type":"ContainerStarted","Data":"2249f7e225a551a90b43d264249769174d7eb251e11aadb3af574a4c1459ea1b"} Nov 29 04:57:03 crc kubenswrapper[4799]: I1129 04:57:03.319827 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"f6a2c6d8-ac5d-4b14-a907-d624f22d5f1b","Type":"ContainerStarted","Data":"c48fe7d528bdd04f9d953a3a0e4e004823948dc947aa5e8a9470f8853dce8144"} Nov 29 04:57:03 crc kubenswrapper[4799]: I1129 04:57:03.320322 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Nov 29 04:57:03 crc kubenswrapper[4799]: I1129 04:57:03.343447 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.3433884750000002 podStartE2EDuration="2.343388475s" podCreationTimestamp="2025-11-29 04:57:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:57:03.340290409 +0000 UTC m=+1098.983220809" watchObservedRunningTime="2025-11-29 04:57:03.343388475 +0000 UTC m=+1098.986318875" Nov 29 04:57:11 crc kubenswrapper[4799]: I1129 04:57:11.745323 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.351242 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-hlvdl"] Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.352900 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-hlvdl" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.359175 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.368902 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-hlvdl"] Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.369972 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.429350 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a28f76a7-54e5-4434-ba25-d3d046b30d5a-config-data\") pod \"nova-cell0-cell-mapping-hlvdl\" (UID: \"a28f76a7-54e5-4434-ba25-d3d046b30d5a\") " pod="openstack/nova-cell0-cell-mapping-hlvdl" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.429599 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a28f76a7-54e5-4434-ba25-d3d046b30d5a-scripts\") pod \"nova-cell0-cell-mapping-hlvdl\" (UID: \"a28f76a7-54e5-4434-ba25-d3d046b30d5a\") " pod="openstack/nova-cell0-cell-mapping-hlvdl" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.429667 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4czhg\" (UniqueName: \"kubernetes.io/projected/a28f76a7-54e5-4434-ba25-d3d046b30d5a-kube-api-access-4czhg\") pod \"nova-cell0-cell-mapping-hlvdl\" (UID: \"a28f76a7-54e5-4434-ba25-d3d046b30d5a\") " pod="openstack/nova-cell0-cell-mapping-hlvdl" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.429950 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a28f76a7-54e5-4434-ba25-d3d046b30d5a-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-hlvdl\" (UID: \"a28f76a7-54e5-4434-ba25-d3d046b30d5a\") " pod="openstack/nova-cell0-cell-mapping-hlvdl" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.531707 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a28f76a7-54e5-4434-ba25-d3d046b30d5a-config-data\") pod \"nova-cell0-cell-mapping-hlvdl\" (UID: \"a28f76a7-54e5-4434-ba25-d3d046b30d5a\") " pod="openstack/nova-cell0-cell-mapping-hlvdl" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.531806 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a28f76a7-54e5-4434-ba25-d3d046b30d5a-scripts\") pod \"nova-cell0-cell-mapping-hlvdl\" (UID: \"a28f76a7-54e5-4434-ba25-d3d046b30d5a\") " pod="openstack/nova-cell0-cell-mapping-hlvdl" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.531833 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4czhg\" (UniqueName: \"kubernetes.io/projected/a28f76a7-54e5-4434-ba25-d3d046b30d5a-kube-api-access-4czhg\") pod \"nova-cell0-cell-mapping-hlvdl\" (UID: \"a28f76a7-54e5-4434-ba25-d3d046b30d5a\") " pod="openstack/nova-cell0-cell-mapping-hlvdl" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.531912 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a28f76a7-54e5-4434-ba25-d3d046b30d5a-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-hlvdl\" (UID: \"a28f76a7-54e5-4434-ba25-d3d046b30d5a\") " pod="openstack/nova-cell0-cell-mapping-hlvdl" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.540596 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a28f76a7-54e5-4434-ba25-d3d046b30d5a-scripts\") pod \"nova-cell0-cell-mapping-hlvdl\" (UID: \"a28f76a7-54e5-4434-ba25-d3d046b30d5a\") " pod="openstack/nova-cell0-cell-mapping-hlvdl" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.541315 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a28f76a7-54e5-4434-ba25-d3d046b30d5a-config-data\") pod \"nova-cell0-cell-mapping-hlvdl\" (UID: \"a28f76a7-54e5-4434-ba25-d3d046b30d5a\") " pod="openstack/nova-cell0-cell-mapping-hlvdl" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.541447 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a28f76a7-54e5-4434-ba25-d3d046b30d5a-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-hlvdl\" (UID: \"a28f76a7-54e5-4434-ba25-d3d046b30d5a\") " pod="openstack/nova-cell0-cell-mapping-hlvdl" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.565844 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4czhg\" (UniqueName: \"kubernetes.io/projected/a28f76a7-54e5-4434-ba25-d3d046b30d5a-kube-api-access-4czhg\") pod \"nova-cell0-cell-mapping-hlvdl\" (UID: \"a28f76a7-54e5-4434-ba25-d3d046b30d5a\") " pod="openstack/nova-cell0-cell-mapping-hlvdl" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.579400 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.580989 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.589207 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.615659 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.675085 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-hlvdl" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.677979 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.682721 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.686833 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.708433 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.738030 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbpw9\" (UniqueName: \"kubernetes.io/projected/3603071d-80f3-4d51-983f-282a0fb33ac4-kube-api-access-bbpw9\") pod \"nova-cell1-novncproxy-0\" (UID: \"3603071d-80f3-4d51-983f-282a0fb33ac4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.738095 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3603071d-80f3-4d51-983f-282a0fb33ac4-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3603071d-80f3-4d51-983f-282a0fb33ac4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.738113 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3603071d-80f3-4d51-983f-282a0fb33ac4-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3603071d-80f3-4d51-983f-282a0fb33ac4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.769406 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.771921 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.786650 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.787170 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.839600 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae86ab9e-2d28-4074-8364-9d1a01d77b62-config-data\") pod \"nova-api-0\" (UID: \"ae86ab9e-2d28-4074-8364-9d1a01d77b62\") " pod="openstack/nova-api-0" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.839649 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8v6rg\" (UniqueName: \"kubernetes.io/projected/ae86ab9e-2d28-4074-8364-9d1a01d77b62-kube-api-access-8v6rg\") pod \"nova-api-0\" (UID: \"ae86ab9e-2d28-4074-8364-9d1a01d77b62\") " pod="openstack/nova-api-0" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.843750 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae86ab9e-2d28-4074-8364-9d1a01d77b62-logs\") pod \"nova-api-0\" (UID: \"ae86ab9e-2d28-4074-8364-9d1a01d77b62\") " pod="openstack/nova-api-0" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.843840 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbpw9\" (UniqueName: \"kubernetes.io/projected/3603071d-80f3-4d51-983f-282a0fb33ac4-kube-api-access-bbpw9\") pod \"nova-cell1-novncproxy-0\" (UID: \"3603071d-80f3-4d51-983f-282a0fb33ac4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.843899 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3603071d-80f3-4d51-983f-282a0fb33ac4-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3603071d-80f3-4d51-983f-282a0fb33ac4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.843930 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3603071d-80f3-4d51-983f-282a0fb33ac4-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3603071d-80f3-4d51-983f-282a0fb33ac4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.844004 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rfvz\" (UniqueName: \"kubernetes.io/projected/06f32a7b-0ded-4927-9b8b-55fc1ecbb296-kube-api-access-8rfvz\") pod \"nova-scheduler-0\" (UID: \"06f32a7b-0ded-4927-9b8b-55fc1ecbb296\") " pod="openstack/nova-scheduler-0" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.844083 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06f32a7b-0ded-4927-9b8b-55fc1ecbb296-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"06f32a7b-0ded-4927-9b8b-55fc1ecbb296\") " pod="openstack/nova-scheduler-0" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.844132 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae86ab9e-2d28-4074-8364-9d1a01d77b62-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ae86ab9e-2d28-4074-8364-9d1a01d77b62\") " pod="openstack/nova-api-0" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.844217 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06f32a7b-0ded-4927-9b8b-55fc1ecbb296-config-data\") pod \"nova-scheduler-0\" (UID: \"06f32a7b-0ded-4927-9b8b-55fc1ecbb296\") " pod="openstack/nova-scheduler-0" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.850244 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3603071d-80f3-4d51-983f-282a0fb33ac4-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3603071d-80f3-4d51-983f-282a0fb33ac4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.853557 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3603071d-80f3-4d51-983f-282a0fb33ac4-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3603071d-80f3-4d51-983f-282a0fb33ac4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.878471 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbpw9\" (UniqueName: \"kubernetes.io/projected/3603071d-80f3-4d51-983f-282a0fb33ac4-kube-api-access-bbpw9\") pod \"nova-cell1-novncproxy-0\" (UID: \"3603071d-80f3-4d51-983f-282a0fb33ac4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.924475 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.926726 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.933266 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.933784 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.946607 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06f32a7b-0ded-4927-9b8b-55fc1ecbb296-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"06f32a7b-0ded-4927-9b8b-55fc1ecbb296\") " pod="openstack/nova-scheduler-0" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.946673 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae86ab9e-2d28-4074-8364-9d1a01d77b62-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ae86ab9e-2d28-4074-8364-9d1a01d77b62\") " pod="openstack/nova-api-0" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.946730 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06f32a7b-0ded-4927-9b8b-55fc1ecbb296-config-data\") pod \"nova-scheduler-0\" (UID: \"06f32a7b-0ded-4927-9b8b-55fc1ecbb296\") " pod="openstack/nova-scheduler-0" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.946821 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae86ab9e-2d28-4074-8364-9d1a01d77b62-config-data\") pod \"nova-api-0\" (UID: \"ae86ab9e-2d28-4074-8364-9d1a01d77b62\") " pod="openstack/nova-api-0" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.946856 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8v6rg\" (UniqueName: \"kubernetes.io/projected/ae86ab9e-2d28-4074-8364-9d1a01d77b62-kube-api-access-8v6rg\") pod \"nova-api-0\" (UID: \"ae86ab9e-2d28-4074-8364-9d1a01d77b62\") " pod="openstack/nova-api-0" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.946912 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae86ab9e-2d28-4074-8364-9d1a01d77b62-logs\") pod \"nova-api-0\" (UID: \"ae86ab9e-2d28-4074-8364-9d1a01d77b62\") " pod="openstack/nova-api-0" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.946980 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rfvz\" (UniqueName: \"kubernetes.io/projected/06f32a7b-0ded-4927-9b8b-55fc1ecbb296-kube-api-access-8rfvz\") pod \"nova-scheduler-0\" (UID: \"06f32a7b-0ded-4927-9b8b-55fc1ecbb296\") " pod="openstack/nova-scheduler-0" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.948048 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae86ab9e-2d28-4074-8364-9d1a01d77b62-logs\") pod \"nova-api-0\" (UID: \"ae86ab9e-2d28-4074-8364-9d1a01d77b62\") " pod="openstack/nova-api-0" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.951634 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae86ab9e-2d28-4074-8364-9d1a01d77b62-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ae86ab9e-2d28-4074-8364-9d1a01d77b62\") " pod="openstack/nova-api-0" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.960839 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.962601 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06f32a7b-0ded-4927-9b8b-55fc1ecbb296-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"06f32a7b-0ded-4927-9b8b-55fc1ecbb296\") " pod="openstack/nova-scheduler-0" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.964145 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae86ab9e-2d28-4074-8364-9d1a01d77b62-config-data\") pod \"nova-api-0\" (UID: \"ae86ab9e-2d28-4074-8364-9d1a01d77b62\") " pod="openstack/nova-api-0" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.968656 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-566b5b7845-8frfb"] Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.970897 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-566b5b7845-8frfb" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.968723 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06f32a7b-0ded-4927-9b8b-55fc1ecbb296-config-data\") pod \"nova-scheduler-0\" (UID: \"06f32a7b-0ded-4927-9b8b-55fc1ecbb296\") " pod="openstack/nova-scheduler-0" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.982971 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rfvz\" (UniqueName: \"kubernetes.io/projected/06f32a7b-0ded-4927-9b8b-55fc1ecbb296-kube-api-access-8rfvz\") pod \"nova-scheduler-0\" (UID: \"06f32a7b-0ded-4927-9b8b-55fc1ecbb296\") " pod="openstack/nova-scheduler-0" Nov 29 04:57:12 crc kubenswrapper[4799]: I1129 04:57:12.985064 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8v6rg\" (UniqueName: \"kubernetes.io/projected/ae86ab9e-2d28-4074-8364-9d1a01d77b62-kube-api-access-8v6rg\") pod \"nova-api-0\" (UID: \"ae86ab9e-2d28-4074-8364-9d1a01d77b62\") " pod="openstack/nova-api-0" Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.002628 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-566b5b7845-8frfb"] Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.009946 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.048628 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68fa593e-4704-40d7-a86d-ae984e5129cc-config-data\") pod \"nova-metadata-0\" (UID: \"68fa593e-4704-40d7-a86d-ae984e5129cc\") " pod="openstack/nova-metadata-0" Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.048697 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88rdm\" (UniqueName: \"kubernetes.io/projected/68fa593e-4704-40d7-a86d-ae984e5129cc-kube-api-access-88rdm\") pod \"nova-metadata-0\" (UID: \"68fa593e-4704-40d7-a86d-ae984e5129cc\") " pod="openstack/nova-metadata-0" Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.048721 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68fa593e-4704-40d7-a86d-ae984e5129cc-logs\") pod \"nova-metadata-0\" (UID: \"68fa593e-4704-40d7-a86d-ae984e5129cc\") " pod="openstack/nova-metadata-0" Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.048845 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjv8g\" (UniqueName: \"kubernetes.io/projected/00ffa619-b06e-46bc-832a-f7fbc2e0efff-kube-api-access-wjv8g\") pod \"dnsmasq-dns-566b5b7845-8frfb\" (UID: \"00ffa619-b06e-46bc-832a-f7fbc2e0efff\") " pod="openstack/dnsmasq-dns-566b5b7845-8frfb" Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.048877 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00ffa619-b06e-46bc-832a-f7fbc2e0efff-config\") pod \"dnsmasq-dns-566b5b7845-8frfb\" (UID: \"00ffa619-b06e-46bc-832a-f7fbc2e0efff\") " pod="openstack/dnsmasq-dns-566b5b7845-8frfb" Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.048915 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68fa593e-4704-40d7-a86d-ae984e5129cc-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"68fa593e-4704-40d7-a86d-ae984e5129cc\") " pod="openstack/nova-metadata-0" Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.048942 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/00ffa619-b06e-46bc-832a-f7fbc2e0efff-ovsdbserver-nb\") pod \"dnsmasq-dns-566b5b7845-8frfb\" (UID: \"00ffa619-b06e-46bc-832a-f7fbc2e0efff\") " pod="openstack/dnsmasq-dns-566b5b7845-8frfb" Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.048971 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00ffa619-b06e-46bc-832a-f7fbc2e0efff-dns-svc\") pod \"dnsmasq-dns-566b5b7845-8frfb\" (UID: \"00ffa619-b06e-46bc-832a-f7fbc2e0efff\") " pod="openstack/dnsmasq-dns-566b5b7845-8frfb" Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.049005 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/00ffa619-b06e-46bc-832a-f7fbc2e0efff-ovsdbserver-sb\") pod \"dnsmasq-dns-566b5b7845-8frfb\" (UID: \"00ffa619-b06e-46bc-832a-f7fbc2e0efff\") " pod="openstack/dnsmasq-dns-566b5b7845-8frfb" Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.118307 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.151090 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68fa593e-4704-40d7-a86d-ae984e5129cc-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"68fa593e-4704-40d7-a86d-ae984e5129cc\") " pod="openstack/nova-metadata-0" Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.151153 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/00ffa619-b06e-46bc-832a-f7fbc2e0efff-ovsdbserver-nb\") pod \"dnsmasq-dns-566b5b7845-8frfb\" (UID: \"00ffa619-b06e-46bc-832a-f7fbc2e0efff\") " pod="openstack/dnsmasq-dns-566b5b7845-8frfb" Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.151187 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00ffa619-b06e-46bc-832a-f7fbc2e0efff-dns-svc\") pod \"dnsmasq-dns-566b5b7845-8frfb\" (UID: \"00ffa619-b06e-46bc-832a-f7fbc2e0efff\") " pod="openstack/dnsmasq-dns-566b5b7845-8frfb" Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.151215 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/00ffa619-b06e-46bc-832a-f7fbc2e0efff-ovsdbserver-sb\") pod \"dnsmasq-dns-566b5b7845-8frfb\" (UID: \"00ffa619-b06e-46bc-832a-f7fbc2e0efff\") " pod="openstack/dnsmasq-dns-566b5b7845-8frfb" Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.151253 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68fa593e-4704-40d7-a86d-ae984e5129cc-config-data\") pod \"nova-metadata-0\" (UID: \"68fa593e-4704-40d7-a86d-ae984e5129cc\") " pod="openstack/nova-metadata-0" Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.158351 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00ffa619-b06e-46bc-832a-f7fbc2e0efff-dns-svc\") pod \"dnsmasq-dns-566b5b7845-8frfb\" (UID: \"00ffa619-b06e-46bc-832a-f7fbc2e0efff\") " pod="openstack/dnsmasq-dns-566b5b7845-8frfb" Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.164831 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/00ffa619-b06e-46bc-832a-f7fbc2e0efff-ovsdbserver-nb\") pod \"dnsmasq-dns-566b5b7845-8frfb\" (UID: \"00ffa619-b06e-46bc-832a-f7fbc2e0efff\") " pod="openstack/dnsmasq-dns-566b5b7845-8frfb" Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.165508 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/00ffa619-b06e-46bc-832a-f7fbc2e0efff-ovsdbserver-sb\") pod \"dnsmasq-dns-566b5b7845-8frfb\" (UID: \"00ffa619-b06e-46bc-832a-f7fbc2e0efff\") " pod="openstack/dnsmasq-dns-566b5b7845-8frfb" Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.157131 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88rdm\" (UniqueName: \"kubernetes.io/projected/68fa593e-4704-40d7-a86d-ae984e5129cc-kube-api-access-88rdm\") pod \"nova-metadata-0\" (UID: \"68fa593e-4704-40d7-a86d-ae984e5129cc\") " pod="openstack/nova-metadata-0" Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.168572 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68fa593e-4704-40d7-a86d-ae984e5129cc-logs\") pod \"nova-metadata-0\" (UID: \"68fa593e-4704-40d7-a86d-ae984e5129cc\") " pod="openstack/nova-metadata-0" Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.168949 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjv8g\" (UniqueName: \"kubernetes.io/projected/00ffa619-b06e-46bc-832a-f7fbc2e0efff-kube-api-access-wjv8g\") pod \"dnsmasq-dns-566b5b7845-8frfb\" (UID: \"00ffa619-b06e-46bc-832a-f7fbc2e0efff\") " pod="openstack/dnsmasq-dns-566b5b7845-8frfb" Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.169053 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00ffa619-b06e-46bc-832a-f7fbc2e0efff-config\") pod \"dnsmasq-dns-566b5b7845-8frfb\" (UID: \"00ffa619-b06e-46bc-832a-f7fbc2e0efff\") " pod="openstack/dnsmasq-dns-566b5b7845-8frfb" Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.170395 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00ffa619-b06e-46bc-832a-f7fbc2e0efff-config\") pod \"dnsmasq-dns-566b5b7845-8frfb\" (UID: \"00ffa619-b06e-46bc-832a-f7fbc2e0efff\") " pod="openstack/dnsmasq-dns-566b5b7845-8frfb" Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.170694 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68fa593e-4704-40d7-a86d-ae984e5129cc-logs\") pod \"nova-metadata-0\" (UID: \"68fa593e-4704-40d7-a86d-ae984e5129cc\") " pod="openstack/nova-metadata-0" Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.197707 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68fa593e-4704-40d7-a86d-ae984e5129cc-config-data\") pod \"nova-metadata-0\" (UID: \"68fa593e-4704-40d7-a86d-ae984e5129cc\") " pod="openstack/nova-metadata-0" Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.198072 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68fa593e-4704-40d7-a86d-ae984e5129cc-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"68fa593e-4704-40d7-a86d-ae984e5129cc\") " pod="openstack/nova-metadata-0" Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.199403 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjv8g\" (UniqueName: \"kubernetes.io/projected/00ffa619-b06e-46bc-832a-f7fbc2e0efff-kube-api-access-wjv8g\") pod \"dnsmasq-dns-566b5b7845-8frfb\" (UID: \"00ffa619-b06e-46bc-832a-f7fbc2e0efff\") " pod="openstack/dnsmasq-dns-566b5b7845-8frfb" Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.200441 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88rdm\" (UniqueName: \"kubernetes.io/projected/68fa593e-4704-40d7-a86d-ae984e5129cc-kube-api-access-88rdm\") pod \"nova-metadata-0\" (UID: \"68fa593e-4704-40d7-a86d-ae984e5129cc\") " pod="openstack/nova-metadata-0" Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.273596 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.300746 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-566b5b7845-8frfb" Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.429888 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-hlvdl"] Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.581719 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-s5l5s"] Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.590483 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-s5l5s" Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.593664 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.593964 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.629378 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-s5l5s"] Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.669611 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.684746 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67b47\" (UniqueName: \"kubernetes.io/projected/302ec8e8-d6a9-4bf2-a586-d6e95fcb232a-kube-api-access-67b47\") pod \"nova-cell1-conductor-db-sync-s5l5s\" (UID: \"302ec8e8-d6a9-4bf2-a586-d6e95fcb232a\") " pod="openstack/nova-cell1-conductor-db-sync-s5l5s" Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.684968 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/302ec8e8-d6a9-4bf2-a586-d6e95fcb232a-config-data\") pod \"nova-cell1-conductor-db-sync-s5l5s\" (UID: \"302ec8e8-d6a9-4bf2-a586-d6e95fcb232a\") " pod="openstack/nova-cell1-conductor-db-sync-s5l5s" Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.685215 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/302ec8e8-d6a9-4bf2-a586-d6e95fcb232a-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-s5l5s\" (UID: \"302ec8e8-d6a9-4bf2-a586-d6e95fcb232a\") " pod="openstack/nova-cell1-conductor-db-sync-s5l5s" Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.685239 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/302ec8e8-d6a9-4bf2-a586-d6e95fcb232a-scripts\") pod \"nova-cell1-conductor-db-sync-s5l5s\" (UID: \"302ec8e8-d6a9-4bf2-a586-d6e95fcb232a\") " pod="openstack/nova-cell1-conductor-db-sync-s5l5s" Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.762737 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 29 04:57:13 crc kubenswrapper[4799]: W1129 04:57:13.776387 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3603071d_80f3_4d51_983f_282a0fb33ac4.slice/crio-2958e40b129c8a386f93b97faedd3136b51b07eae64ea53be56627f207f43e93 WatchSource:0}: Error finding container 2958e40b129c8a386f93b97faedd3136b51b07eae64ea53be56627f207f43e93: Status 404 returned error can't find the container with id 2958e40b129c8a386f93b97faedd3136b51b07eae64ea53be56627f207f43e93 Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.786518 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/302ec8e8-d6a9-4bf2-a586-d6e95fcb232a-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-s5l5s\" (UID: \"302ec8e8-d6a9-4bf2-a586-d6e95fcb232a\") " pod="openstack/nova-cell1-conductor-db-sync-s5l5s" Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.786573 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/302ec8e8-d6a9-4bf2-a586-d6e95fcb232a-scripts\") pod \"nova-cell1-conductor-db-sync-s5l5s\" (UID: \"302ec8e8-d6a9-4bf2-a586-d6e95fcb232a\") " pod="openstack/nova-cell1-conductor-db-sync-s5l5s" Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.786653 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67b47\" (UniqueName: \"kubernetes.io/projected/302ec8e8-d6a9-4bf2-a586-d6e95fcb232a-kube-api-access-67b47\") pod \"nova-cell1-conductor-db-sync-s5l5s\" (UID: \"302ec8e8-d6a9-4bf2-a586-d6e95fcb232a\") " pod="openstack/nova-cell1-conductor-db-sync-s5l5s" Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.786772 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/302ec8e8-d6a9-4bf2-a586-d6e95fcb232a-config-data\") pod \"nova-cell1-conductor-db-sync-s5l5s\" (UID: \"302ec8e8-d6a9-4bf2-a586-d6e95fcb232a\") " pod="openstack/nova-cell1-conductor-db-sync-s5l5s" Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.795554 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/302ec8e8-d6a9-4bf2-a586-d6e95fcb232a-scripts\") pod \"nova-cell1-conductor-db-sync-s5l5s\" (UID: \"302ec8e8-d6a9-4bf2-a586-d6e95fcb232a\") " pod="openstack/nova-cell1-conductor-db-sync-s5l5s" Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.795599 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/302ec8e8-d6a9-4bf2-a586-d6e95fcb232a-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-s5l5s\" (UID: \"302ec8e8-d6a9-4bf2-a586-d6e95fcb232a\") " pod="openstack/nova-cell1-conductor-db-sync-s5l5s" Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.795674 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/302ec8e8-d6a9-4bf2-a586-d6e95fcb232a-config-data\") pod \"nova-cell1-conductor-db-sync-s5l5s\" (UID: \"302ec8e8-d6a9-4bf2-a586-d6e95fcb232a\") " pod="openstack/nova-cell1-conductor-db-sync-s5l5s" Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.811189 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67b47\" (UniqueName: \"kubernetes.io/projected/302ec8e8-d6a9-4bf2-a586-d6e95fcb232a-kube-api-access-67b47\") pod \"nova-cell1-conductor-db-sync-s5l5s\" (UID: \"302ec8e8-d6a9-4bf2-a586-d6e95fcb232a\") " pod="openstack/nova-cell1-conductor-db-sync-s5l5s" Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.830570 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-s5l5s" Nov 29 04:57:13 crc kubenswrapper[4799]: I1129 04:57:13.990705 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 29 04:57:14 crc kubenswrapper[4799]: I1129 04:57:14.019451 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 29 04:57:14 crc kubenswrapper[4799]: I1129 04:57:14.169984 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-566b5b7845-8frfb"] Nov 29 04:57:14 crc kubenswrapper[4799]: I1129 04:57:14.415375 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-s5l5s"] Nov 29 04:57:14 crc kubenswrapper[4799]: W1129 04:57:14.424008 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod302ec8e8_d6a9_4bf2_a586_d6e95fcb232a.slice/crio-4bb6de749e7dcbf61ba7e6655870209b24923730cbc44cce6cf6b6ed2b0ef388 WatchSource:0}: Error finding container 4bb6de749e7dcbf61ba7e6655870209b24923730cbc44cce6cf6b6ed2b0ef388: Status 404 returned error can't find the container with id 4bb6de749e7dcbf61ba7e6655870209b24923730cbc44cce6cf6b6ed2b0ef388 Nov 29 04:57:14 crc kubenswrapper[4799]: I1129 04:57:14.483540 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-hlvdl" event={"ID":"a28f76a7-54e5-4434-ba25-d3d046b30d5a","Type":"ContainerStarted","Data":"9e2c281a078f1e86deb92e6cc370f3da6dbb46757f202dd36faf43f9f9471e25"} Nov 29 04:57:14 crc kubenswrapper[4799]: I1129 04:57:14.483611 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-hlvdl" event={"ID":"a28f76a7-54e5-4434-ba25-d3d046b30d5a","Type":"ContainerStarted","Data":"7a02375c1c673ae40e9efcab03118a1d57fe651613ed3023167240d74aab83b3"} Nov 29 04:57:14 crc kubenswrapper[4799]: I1129 04:57:14.486471 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ae86ab9e-2d28-4074-8364-9d1a01d77b62","Type":"ContainerStarted","Data":"a9201a8933663eb523fc944cb23e7770003984dbf12f075f256fa00ead0f1ec6"} Nov 29 04:57:14 crc kubenswrapper[4799]: I1129 04:57:14.488154 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3603071d-80f3-4d51-983f-282a0fb33ac4","Type":"ContainerStarted","Data":"2958e40b129c8a386f93b97faedd3136b51b07eae64ea53be56627f207f43e93"} Nov 29 04:57:14 crc kubenswrapper[4799]: I1129 04:57:14.489752 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"68fa593e-4704-40d7-a86d-ae984e5129cc","Type":"ContainerStarted","Data":"18ff036b3ca2b7c2caa69921456d2aafa72c8241293c8f4e760ad8718d058a19"} Nov 29 04:57:14 crc kubenswrapper[4799]: I1129 04:57:14.491084 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"06f32a7b-0ded-4927-9b8b-55fc1ecbb296","Type":"ContainerStarted","Data":"31b73efb28ab805bbabc3c8844bf23a0e17e766a2449efd7457674748c96c357"} Nov 29 04:57:14 crc kubenswrapper[4799]: I1129 04:57:14.493325 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-s5l5s" event={"ID":"302ec8e8-d6a9-4bf2-a586-d6e95fcb232a","Type":"ContainerStarted","Data":"4bb6de749e7dcbf61ba7e6655870209b24923730cbc44cce6cf6b6ed2b0ef388"} Nov 29 04:57:14 crc kubenswrapper[4799]: I1129 04:57:14.503035 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-566b5b7845-8frfb" event={"ID":"00ffa619-b06e-46bc-832a-f7fbc2e0efff","Type":"ContainerStarted","Data":"6697d3b3a1c48517b7807ce467aaca6e6447664573c9533119068982bbde9b55"} Nov 29 04:57:14 crc kubenswrapper[4799]: I1129 04:57:14.519151 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-hlvdl" podStartSLOduration=2.519124185 podStartE2EDuration="2.519124185s" podCreationTimestamp="2025-11-29 04:57:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:57:14.503807061 +0000 UTC m=+1110.146737461" watchObservedRunningTime="2025-11-29 04:57:14.519124185 +0000 UTC m=+1110.162054585" Nov 29 04:57:15 crc kubenswrapper[4799]: I1129 04:57:15.527305 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-s5l5s" event={"ID":"302ec8e8-d6a9-4bf2-a586-d6e95fcb232a","Type":"ContainerStarted","Data":"405d2216d4e4ae4783cbddda45b5628a8388149b26b0386cbeb616ebef54dc2d"} Nov 29 04:57:15 crc kubenswrapper[4799]: I1129 04:57:15.529866 4799 generic.go:334] "Generic (PLEG): container finished" podID="00ffa619-b06e-46bc-832a-f7fbc2e0efff" containerID="51eb9bd77778fbf6140e3e6018fb61b2ac6ff0ca188c70a433822c521d2a1908" exitCode=0 Nov 29 04:57:15 crc kubenswrapper[4799]: I1129 04:57:15.529985 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-566b5b7845-8frfb" event={"ID":"00ffa619-b06e-46bc-832a-f7fbc2e0efff","Type":"ContainerDied","Data":"51eb9bd77778fbf6140e3e6018fb61b2ac6ff0ca188c70a433822c521d2a1908"} Nov 29 04:57:15 crc kubenswrapper[4799]: I1129 04:57:15.547741 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-s5l5s" podStartSLOduration=2.547701403 podStartE2EDuration="2.547701403s" podCreationTimestamp="2025-11-29 04:57:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:57:15.54336224 +0000 UTC m=+1111.186292640" watchObservedRunningTime="2025-11-29 04:57:15.547701403 +0000 UTC m=+1111.190631804" Nov 29 04:57:16 crc kubenswrapper[4799]: I1129 04:57:16.143680 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 29 04:57:16 crc kubenswrapper[4799]: I1129 04:57:16.164581 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 29 04:57:18 crc kubenswrapper[4799]: I1129 04:57:18.812437 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 29 04:57:19 crc kubenswrapper[4799]: I1129 04:57:19.576016 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"68fa593e-4704-40d7-a86d-ae984e5129cc","Type":"ContainerStarted","Data":"cfeb3da10847f7ebeda995678cedbbba4c0edf34e678a1a155acb97431284903"} Nov 29 04:57:19 crc kubenswrapper[4799]: I1129 04:57:19.577072 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"68fa593e-4704-40d7-a86d-ae984e5129cc","Type":"ContainerStarted","Data":"003f5f407757a793f711e60be6c4870f1a4a1e1cbe2647fdf10cc619f51a2e01"} Nov 29 04:57:19 crc kubenswrapper[4799]: I1129 04:57:19.576189 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="68fa593e-4704-40d7-a86d-ae984e5129cc" containerName="nova-metadata-log" containerID="cri-o://003f5f407757a793f711e60be6c4870f1a4a1e1cbe2647fdf10cc619f51a2e01" gracePeriod=30 Nov 29 04:57:19 crc kubenswrapper[4799]: I1129 04:57:19.576509 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="68fa593e-4704-40d7-a86d-ae984e5129cc" containerName="nova-metadata-metadata" containerID="cri-o://cfeb3da10847f7ebeda995678cedbbba4c0edf34e678a1a155acb97431284903" gracePeriod=30 Nov 29 04:57:19 crc kubenswrapper[4799]: I1129 04:57:19.578474 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"06f32a7b-0ded-4927-9b8b-55fc1ecbb296","Type":"ContainerStarted","Data":"eb674ecc4ca4399e936f237d941a66c8354d24c9fb184e868257e4e51dca7f89"} Nov 29 04:57:19 crc kubenswrapper[4799]: I1129 04:57:19.582370 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-566b5b7845-8frfb" event={"ID":"00ffa619-b06e-46bc-832a-f7fbc2e0efff","Type":"ContainerStarted","Data":"cd2b609ab12cb328932676754756fc40228b18f8b81ffb0ef43d61eb2664f941"} Nov 29 04:57:19 crc kubenswrapper[4799]: I1129 04:57:19.582586 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-566b5b7845-8frfb" Nov 29 04:57:19 crc kubenswrapper[4799]: I1129 04:57:19.591931 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ae86ab9e-2d28-4074-8364-9d1a01d77b62","Type":"ContainerStarted","Data":"ac913cd6f6b3c6cd2d97df30c634ac594219e8c013eb65754ed899e61dbaf202"} Nov 29 04:57:19 crc kubenswrapper[4799]: I1129 04:57:19.591987 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ae86ab9e-2d28-4074-8364-9d1a01d77b62","Type":"ContainerStarted","Data":"c1d81c8095f715598cc9fb838de2ce04ef7ca5c82f6b28b2c88ab54e618a9fea"} Nov 29 04:57:19 crc kubenswrapper[4799]: I1129 04:57:19.595766 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3603071d-80f3-4d51-983f-282a0fb33ac4","Type":"ContainerStarted","Data":"95d7d63500012b1fa8cd80b35d7674be5e6fe32e79244a1d378e9daf4ce43303"} Nov 29 04:57:19 crc kubenswrapper[4799]: I1129 04:57:19.596029 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="3603071d-80f3-4d51-983f-282a0fb33ac4" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://95d7d63500012b1fa8cd80b35d7674be5e6fe32e79244a1d378e9daf4ce43303" gracePeriod=30 Nov 29 04:57:19 crc kubenswrapper[4799]: I1129 04:57:19.605437 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.714868423 podStartE2EDuration="7.605418857s" podCreationTimestamp="2025-11-29 04:57:12 +0000 UTC" firstStartedPulling="2025-11-29 04:57:14.019883215 +0000 UTC m=+1109.662813615" lastFinishedPulling="2025-11-29 04:57:18.910433649 +0000 UTC m=+1114.553364049" observedRunningTime="2025-11-29 04:57:19.601826131 +0000 UTC m=+1115.244756531" watchObservedRunningTime="2025-11-29 04:57:19.605418857 +0000 UTC m=+1115.248349257" Nov 29 04:57:19 crc kubenswrapper[4799]: I1129 04:57:19.626156 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.493464735 podStartE2EDuration="7.62610679s" podCreationTimestamp="2025-11-29 04:57:12 +0000 UTC" firstStartedPulling="2025-11-29 04:57:13.778107451 +0000 UTC m=+1109.421037851" lastFinishedPulling="2025-11-29 04:57:18.910749506 +0000 UTC m=+1114.553679906" observedRunningTime="2025-11-29 04:57:19.623802064 +0000 UTC m=+1115.266732474" watchObservedRunningTime="2025-11-29 04:57:19.62610679 +0000 UTC m=+1115.269037230" Nov 29 04:57:19 crc kubenswrapper[4799]: I1129 04:57:19.653975 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.439505101 podStartE2EDuration="7.653951632s" podCreationTimestamp="2025-11-29 04:57:12 +0000 UTC" firstStartedPulling="2025-11-29 04:57:13.683289145 +0000 UTC m=+1109.326219545" lastFinishedPulling="2025-11-29 04:57:18.897735676 +0000 UTC m=+1114.540666076" observedRunningTime="2025-11-29 04:57:19.645977092 +0000 UTC m=+1115.288907502" watchObservedRunningTime="2025-11-29 04:57:19.653951632 +0000 UTC m=+1115.296882032" Nov 29 04:57:19 crc kubenswrapper[4799]: I1129 04:57:19.674106 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.788790614 podStartE2EDuration="7.674079031s" podCreationTimestamp="2025-11-29 04:57:12 +0000 UTC" firstStartedPulling="2025-11-29 04:57:14.011538997 +0000 UTC m=+1109.654469397" lastFinishedPulling="2025-11-29 04:57:18.896827404 +0000 UTC m=+1114.539757814" observedRunningTime="2025-11-29 04:57:19.66815067 +0000 UTC m=+1115.311081080" watchObservedRunningTime="2025-11-29 04:57:19.674079031 +0000 UTC m=+1115.317009431" Nov 29 04:57:19 crc kubenswrapper[4799]: I1129 04:57:19.695516 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-566b5b7845-8frfb" podStartSLOduration=7.69549558 podStartE2EDuration="7.69549558s" podCreationTimestamp="2025-11-29 04:57:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:57:19.691094206 +0000 UTC m=+1115.334024596" watchObservedRunningTime="2025-11-29 04:57:19.69549558 +0000 UTC m=+1115.338425980" Nov 29 04:57:20 crc kubenswrapper[4799]: I1129 04:57:20.609478 4799 generic.go:334] "Generic (PLEG): container finished" podID="68fa593e-4704-40d7-a86d-ae984e5129cc" containerID="003f5f407757a793f711e60be6c4870f1a4a1e1cbe2647fdf10cc619f51a2e01" exitCode=143 Nov 29 04:57:20 crc kubenswrapper[4799]: I1129 04:57:20.609585 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"68fa593e-4704-40d7-a86d-ae984e5129cc","Type":"ContainerDied","Data":"003f5f407757a793f711e60be6c4870f1a4a1e1cbe2647fdf10cc619f51a2e01"} Nov 29 04:57:21 crc kubenswrapper[4799]: I1129 04:57:21.832623 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 29 04:57:21 crc kubenswrapper[4799]: I1129 04:57:21.832907 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="9c406f33-9e26-4226-b6dd-c7449fa96b9f" containerName="kube-state-metrics" containerID="cri-o://ab9bfe5165025b9af4a628e94c467832cbbce81cb0f59b89fbbe2bd8f224ef8d" gracePeriod=30 Nov 29 04:57:22 crc kubenswrapper[4799]: I1129 04:57:22.421367 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 29 04:57:22 crc kubenswrapper[4799]: I1129 04:57:22.548647 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-77k8h\" (UniqueName: \"kubernetes.io/projected/9c406f33-9e26-4226-b6dd-c7449fa96b9f-kube-api-access-77k8h\") pod \"9c406f33-9e26-4226-b6dd-c7449fa96b9f\" (UID: \"9c406f33-9e26-4226-b6dd-c7449fa96b9f\") " Nov 29 04:57:22 crc kubenswrapper[4799]: I1129 04:57:22.561391 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c406f33-9e26-4226-b6dd-c7449fa96b9f-kube-api-access-77k8h" (OuterVolumeSpecName: "kube-api-access-77k8h") pod "9c406f33-9e26-4226-b6dd-c7449fa96b9f" (UID: "9c406f33-9e26-4226-b6dd-c7449fa96b9f"). InnerVolumeSpecName "kube-api-access-77k8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:57:22 crc kubenswrapper[4799]: I1129 04:57:22.634202 4799 generic.go:334] "Generic (PLEG): container finished" podID="a28f76a7-54e5-4434-ba25-d3d046b30d5a" containerID="9e2c281a078f1e86deb92e6cc370f3da6dbb46757f202dd36faf43f9f9471e25" exitCode=0 Nov 29 04:57:22 crc kubenswrapper[4799]: I1129 04:57:22.634292 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-hlvdl" event={"ID":"a28f76a7-54e5-4434-ba25-d3d046b30d5a","Type":"ContainerDied","Data":"9e2c281a078f1e86deb92e6cc370f3da6dbb46757f202dd36faf43f9f9471e25"} Nov 29 04:57:22 crc kubenswrapper[4799]: I1129 04:57:22.636876 4799 generic.go:334] "Generic (PLEG): container finished" podID="9c406f33-9e26-4226-b6dd-c7449fa96b9f" containerID="ab9bfe5165025b9af4a628e94c467832cbbce81cb0f59b89fbbe2bd8f224ef8d" exitCode=2 Nov 29 04:57:22 crc kubenswrapper[4799]: I1129 04:57:22.636933 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 29 04:57:22 crc kubenswrapper[4799]: I1129 04:57:22.636937 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"9c406f33-9e26-4226-b6dd-c7449fa96b9f","Type":"ContainerDied","Data":"ab9bfe5165025b9af4a628e94c467832cbbce81cb0f59b89fbbe2bd8f224ef8d"} Nov 29 04:57:22 crc kubenswrapper[4799]: I1129 04:57:22.636987 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"9c406f33-9e26-4226-b6dd-c7449fa96b9f","Type":"ContainerDied","Data":"8d2c691761531ac2ef5a5c98cb36d8c73b2daef63c1d86a04566795ca6be7493"} Nov 29 04:57:22 crc kubenswrapper[4799]: I1129 04:57:22.637016 4799 scope.go:117] "RemoveContainer" containerID="ab9bfe5165025b9af4a628e94c467832cbbce81cb0f59b89fbbe2bd8f224ef8d" Nov 29 04:57:22 crc kubenswrapper[4799]: I1129 04:57:22.651872 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-77k8h\" (UniqueName: \"kubernetes.io/projected/9c406f33-9e26-4226-b6dd-c7449fa96b9f-kube-api-access-77k8h\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:22 crc kubenswrapper[4799]: I1129 04:57:22.690745 4799 scope.go:117] "RemoveContainer" containerID="ab9bfe5165025b9af4a628e94c467832cbbce81cb0f59b89fbbe2bd8f224ef8d" Nov 29 04:57:22 crc kubenswrapper[4799]: I1129 04:57:22.696271 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 29 04:57:22 crc kubenswrapper[4799]: E1129 04:57:22.697635 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab9bfe5165025b9af4a628e94c467832cbbce81cb0f59b89fbbe2bd8f224ef8d\": container with ID starting with ab9bfe5165025b9af4a628e94c467832cbbce81cb0f59b89fbbe2bd8f224ef8d not found: ID does not exist" containerID="ab9bfe5165025b9af4a628e94c467832cbbce81cb0f59b89fbbe2bd8f224ef8d" Nov 29 04:57:22 crc kubenswrapper[4799]: I1129 04:57:22.697782 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab9bfe5165025b9af4a628e94c467832cbbce81cb0f59b89fbbe2bd8f224ef8d"} err="failed to get container status \"ab9bfe5165025b9af4a628e94c467832cbbce81cb0f59b89fbbe2bd8f224ef8d\": rpc error: code = NotFound desc = could not find container \"ab9bfe5165025b9af4a628e94c467832cbbce81cb0f59b89fbbe2bd8f224ef8d\": container with ID starting with ab9bfe5165025b9af4a628e94c467832cbbce81cb0f59b89fbbe2bd8f224ef8d not found: ID does not exist" Nov 29 04:57:22 crc kubenswrapper[4799]: I1129 04:57:22.710702 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 29 04:57:22 crc kubenswrapper[4799]: I1129 04:57:22.722778 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 29 04:57:22 crc kubenswrapper[4799]: E1129 04:57:22.723457 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c406f33-9e26-4226-b6dd-c7449fa96b9f" containerName="kube-state-metrics" Nov 29 04:57:22 crc kubenswrapper[4799]: I1129 04:57:22.723484 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c406f33-9e26-4226-b6dd-c7449fa96b9f" containerName="kube-state-metrics" Nov 29 04:57:22 crc kubenswrapper[4799]: I1129 04:57:22.723728 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c406f33-9e26-4226-b6dd-c7449fa96b9f" containerName="kube-state-metrics" Nov 29 04:57:22 crc kubenswrapper[4799]: I1129 04:57:22.724691 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 29 04:57:22 crc kubenswrapper[4799]: I1129 04:57:22.731482 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 29 04:57:22 crc kubenswrapper[4799]: I1129 04:57:22.763719 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Nov 29 04:57:22 crc kubenswrapper[4799]: I1129 04:57:22.763959 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Nov 29 04:57:22 crc kubenswrapper[4799]: I1129 04:57:22.863623 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/66c61a7f-a655-46e9-8a85-de7ad52bfe6f-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"66c61a7f-a655-46e9-8a85-de7ad52bfe6f\") " pod="openstack/kube-state-metrics-0" Nov 29 04:57:22 crc kubenswrapper[4799]: I1129 04:57:22.863725 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66c61a7f-a655-46e9-8a85-de7ad52bfe6f-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"66c61a7f-a655-46e9-8a85-de7ad52bfe6f\") " pod="openstack/kube-state-metrics-0" Nov 29 04:57:22 crc kubenswrapper[4799]: I1129 04:57:22.863870 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lr9fz\" (UniqueName: \"kubernetes.io/projected/66c61a7f-a655-46e9-8a85-de7ad52bfe6f-kube-api-access-lr9fz\") pod \"kube-state-metrics-0\" (UID: \"66c61a7f-a655-46e9-8a85-de7ad52bfe6f\") " pod="openstack/kube-state-metrics-0" Nov 29 04:57:22 crc kubenswrapper[4799]: I1129 04:57:22.863928 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/66c61a7f-a655-46e9-8a85-de7ad52bfe6f-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"66c61a7f-a655-46e9-8a85-de7ad52bfe6f\") " pod="openstack/kube-state-metrics-0" Nov 29 04:57:22 crc kubenswrapper[4799]: I1129 04:57:22.938521 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 29 04:57:22 crc kubenswrapper[4799]: I1129 04:57:22.938938 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d4741f22-658c-40af-bba0-b36c491be636" containerName="ceilometer-central-agent" containerID="cri-o://cb2cd9b60acdc3381d394780aa0fc5cf6c8d8b1a94029038393ac04292fbfb98" gracePeriod=30 Nov 29 04:57:22 crc kubenswrapper[4799]: I1129 04:57:22.939055 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d4741f22-658c-40af-bba0-b36c491be636" containerName="proxy-httpd" containerID="cri-o://d67f65d1c9b133f52aae7c40e3e167681a53206dbdf3f6cf9575afc01f64f4b6" gracePeriod=30 Nov 29 04:57:22 crc kubenswrapper[4799]: I1129 04:57:22.939131 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d4741f22-658c-40af-bba0-b36c491be636" containerName="sg-core" containerID="cri-o://bc58c8ff01459001f39e28a7979a9c4bfd0e961f421845b23dd2cff71163ceb3" gracePeriod=30 Nov 29 04:57:22 crc kubenswrapper[4799]: I1129 04:57:22.939224 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d4741f22-658c-40af-bba0-b36c491be636" containerName="ceilometer-notification-agent" containerID="cri-o://ff92573600da0847b820dabc90979d71078d06a62fe0d43ca46e4ca0b36f4328" gracePeriod=30 Nov 29 04:57:22 crc kubenswrapper[4799]: I1129 04:57:22.962400 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 29 04:57:22 crc kubenswrapper[4799]: I1129 04:57:22.966364 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66c61a7f-a655-46e9-8a85-de7ad52bfe6f-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"66c61a7f-a655-46e9-8a85-de7ad52bfe6f\") " pod="openstack/kube-state-metrics-0" Nov 29 04:57:22 crc kubenswrapper[4799]: I1129 04:57:22.966479 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lr9fz\" (UniqueName: \"kubernetes.io/projected/66c61a7f-a655-46e9-8a85-de7ad52bfe6f-kube-api-access-lr9fz\") pod \"kube-state-metrics-0\" (UID: \"66c61a7f-a655-46e9-8a85-de7ad52bfe6f\") " pod="openstack/kube-state-metrics-0" Nov 29 04:57:22 crc kubenswrapper[4799]: I1129 04:57:22.966525 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/66c61a7f-a655-46e9-8a85-de7ad52bfe6f-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"66c61a7f-a655-46e9-8a85-de7ad52bfe6f\") " pod="openstack/kube-state-metrics-0" Nov 29 04:57:22 crc kubenswrapper[4799]: I1129 04:57:22.966566 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/66c61a7f-a655-46e9-8a85-de7ad52bfe6f-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"66c61a7f-a655-46e9-8a85-de7ad52bfe6f\") " pod="openstack/kube-state-metrics-0" Nov 29 04:57:22 crc kubenswrapper[4799]: I1129 04:57:22.971233 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/66c61a7f-a655-46e9-8a85-de7ad52bfe6f-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"66c61a7f-a655-46e9-8a85-de7ad52bfe6f\") " pod="openstack/kube-state-metrics-0" Nov 29 04:57:22 crc kubenswrapper[4799]: I1129 04:57:22.971592 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66c61a7f-a655-46e9-8a85-de7ad52bfe6f-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"66c61a7f-a655-46e9-8a85-de7ad52bfe6f\") " pod="openstack/kube-state-metrics-0" Nov 29 04:57:22 crc kubenswrapper[4799]: I1129 04:57:22.979558 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/66c61a7f-a655-46e9-8a85-de7ad52bfe6f-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"66c61a7f-a655-46e9-8a85-de7ad52bfe6f\") " pod="openstack/kube-state-metrics-0" Nov 29 04:57:22 crc kubenswrapper[4799]: I1129 04:57:22.988491 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lr9fz\" (UniqueName: \"kubernetes.io/projected/66c61a7f-a655-46e9-8a85-de7ad52bfe6f-kube-api-access-lr9fz\") pod \"kube-state-metrics-0\" (UID: \"66c61a7f-a655-46e9-8a85-de7ad52bfe6f\") " pod="openstack/kube-state-metrics-0" Nov 29 04:57:23 crc kubenswrapper[4799]: I1129 04:57:23.010899 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 29 04:57:23 crc kubenswrapper[4799]: I1129 04:57:23.011302 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 29 04:57:23 crc kubenswrapper[4799]: I1129 04:57:23.043387 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 29 04:57:23 crc kubenswrapper[4799]: I1129 04:57:23.090485 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 29 04:57:23 crc kubenswrapper[4799]: I1129 04:57:23.119704 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 29 04:57:23 crc kubenswrapper[4799]: I1129 04:57:23.120318 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 29 04:57:23 crc kubenswrapper[4799]: I1129 04:57:23.274092 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 29 04:57:23 crc kubenswrapper[4799]: I1129 04:57:23.275600 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 29 04:57:23 crc kubenswrapper[4799]: I1129 04:57:23.602804 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 29 04:57:23 crc kubenswrapper[4799]: I1129 04:57:23.653548 4799 generic.go:334] "Generic (PLEG): container finished" podID="302ec8e8-d6a9-4bf2-a586-d6e95fcb232a" containerID="405d2216d4e4ae4783cbddda45b5628a8388149b26b0386cbeb616ebef54dc2d" exitCode=0 Nov 29 04:57:23 crc kubenswrapper[4799]: I1129 04:57:23.653680 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-s5l5s" event={"ID":"302ec8e8-d6a9-4bf2-a586-d6e95fcb232a","Type":"ContainerDied","Data":"405d2216d4e4ae4783cbddda45b5628a8388149b26b0386cbeb616ebef54dc2d"} Nov 29 04:57:23 crc kubenswrapper[4799]: W1129 04:57:23.656469 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod66c61a7f_a655_46e9_8a85_de7ad52bfe6f.slice/crio-33bc1604771310b97be652234b9b2497cdbcc0d8505ab90cd91f4bb7205e2a72 WatchSource:0}: Error finding container 33bc1604771310b97be652234b9b2497cdbcc0d8505ab90cd91f4bb7205e2a72: Status 404 returned error can't find the container with id 33bc1604771310b97be652234b9b2497cdbcc0d8505ab90cd91f4bb7205e2a72 Nov 29 04:57:23 crc kubenswrapper[4799]: I1129 04:57:23.678878 4799 generic.go:334] "Generic (PLEG): container finished" podID="d4741f22-658c-40af-bba0-b36c491be636" containerID="d67f65d1c9b133f52aae7c40e3e167681a53206dbdf3f6cf9575afc01f64f4b6" exitCode=0 Nov 29 04:57:23 crc kubenswrapper[4799]: I1129 04:57:23.678914 4799 generic.go:334] "Generic (PLEG): container finished" podID="d4741f22-658c-40af-bba0-b36c491be636" containerID="bc58c8ff01459001f39e28a7979a9c4bfd0e961f421845b23dd2cff71163ceb3" exitCode=2 Nov 29 04:57:23 crc kubenswrapper[4799]: I1129 04:57:23.678987 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d4741f22-658c-40af-bba0-b36c491be636","Type":"ContainerDied","Data":"d67f65d1c9b133f52aae7c40e3e167681a53206dbdf3f6cf9575afc01f64f4b6"} Nov 29 04:57:23 crc kubenswrapper[4799]: I1129 04:57:23.679048 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d4741f22-658c-40af-bba0-b36c491be636","Type":"ContainerDied","Data":"bc58c8ff01459001f39e28a7979a9c4bfd0e961f421845b23dd2cff71163ceb3"} Nov 29 04:57:23 crc kubenswrapper[4799]: I1129 04:57:23.723167 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 29 04:57:24 crc kubenswrapper[4799]: I1129 04:57:24.209057 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ae86ab9e-2d28-4074-8364-9d1a01d77b62" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.168:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 29 04:57:24 crc kubenswrapper[4799]: I1129 04:57:24.211142 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ae86ab9e-2d28-4074-8364-9d1a01d77b62" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.168:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 29 04:57:24 crc kubenswrapper[4799]: I1129 04:57:24.276810 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-hlvdl" Nov 29 04:57:24 crc kubenswrapper[4799]: I1129 04:57:24.408572 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a28f76a7-54e5-4434-ba25-d3d046b30d5a-combined-ca-bundle\") pod \"a28f76a7-54e5-4434-ba25-d3d046b30d5a\" (UID: \"a28f76a7-54e5-4434-ba25-d3d046b30d5a\") " Nov 29 04:57:24 crc kubenswrapper[4799]: I1129 04:57:24.408960 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4czhg\" (UniqueName: \"kubernetes.io/projected/a28f76a7-54e5-4434-ba25-d3d046b30d5a-kube-api-access-4czhg\") pod \"a28f76a7-54e5-4434-ba25-d3d046b30d5a\" (UID: \"a28f76a7-54e5-4434-ba25-d3d046b30d5a\") " Nov 29 04:57:24 crc kubenswrapper[4799]: I1129 04:57:24.409184 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a28f76a7-54e5-4434-ba25-d3d046b30d5a-config-data\") pod \"a28f76a7-54e5-4434-ba25-d3d046b30d5a\" (UID: \"a28f76a7-54e5-4434-ba25-d3d046b30d5a\") " Nov 29 04:57:24 crc kubenswrapper[4799]: I1129 04:57:24.409293 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a28f76a7-54e5-4434-ba25-d3d046b30d5a-scripts\") pod \"a28f76a7-54e5-4434-ba25-d3d046b30d5a\" (UID: \"a28f76a7-54e5-4434-ba25-d3d046b30d5a\") " Nov 29 04:57:24 crc kubenswrapper[4799]: I1129 04:57:24.413745 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a28f76a7-54e5-4434-ba25-d3d046b30d5a-scripts" (OuterVolumeSpecName: "scripts") pod "a28f76a7-54e5-4434-ba25-d3d046b30d5a" (UID: "a28f76a7-54e5-4434-ba25-d3d046b30d5a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:57:24 crc kubenswrapper[4799]: I1129 04:57:24.418066 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a28f76a7-54e5-4434-ba25-d3d046b30d5a-kube-api-access-4czhg" (OuterVolumeSpecName: "kube-api-access-4czhg") pod "a28f76a7-54e5-4434-ba25-d3d046b30d5a" (UID: "a28f76a7-54e5-4434-ba25-d3d046b30d5a"). InnerVolumeSpecName "kube-api-access-4czhg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:57:24 crc kubenswrapper[4799]: I1129 04:57:24.439711 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a28f76a7-54e5-4434-ba25-d3d046b30d5a-config-data" (OuterVolumeSpecName: "config-data") pod "a28f76a7-54e5-4434-ba25-d3d046b30d5a" (UID: "a28f76a7-54e5-4434-ba25-d3d046b30d5a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:57:24 crc kubenswrapper[4799]: I1129 04:57:24.447747 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a28f76a7-54e5-4434-ba25-d3d046b30d5a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a28f76a7-54e5-4434-ba25-d3d046b30d5a" (UID: "a28f76a7-54e5-4434-ba25-d3d046b30d5a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:57:24 crc kubenswrapper[4799]: I1129 04:57:24.512102 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a28f76a7-54e5-4434-ba25-d3d046b30d5a-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:24 crc kubenswrapper[4799]: I1129 04:57:24.512148 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a28f76a7-54e5-4434-ba25-d3d046b30d5a-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:24 crc kubenswrapper[4799]: I1129 04:57:24.512158 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a28f76a7-54e5-4434-ba25-d3d046b30d5a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:24 crc kubenswrapper[4799]: I1129 04:57:24.512171 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4czhg\" (UniqueName: \"kubernetes.io/projected/a28f76a7-54e5-4434-ba25-d3d046b30d5a-kube-api-access-4czhg\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:24 crc kubenswrapper[4799]: I1129 04:57:24.679008 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c406f33-9e26-4226-b6dd-c7449fa96b9f" path="/var/lib/kubelet/pods/9c406f33-9e26-4226-b6dd-c7449fa96b9f/volumes" Nov 29 04:57:24 crc kubenswrapper[4799]: I1129 04:57:24.708264 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-hlvdl" Nov 29 04:57:24 crc kubenswrapper[4799]: I1129 04:57:24.710545 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-hlvdl" event={"ID":"a28f76a7-54e5-4434-ba25-d3d046b30d5a","Type":"ContainerDied","Data":"7a02375c1c673ae40e9efcab03118a1d57fe651613ed3023167240d74aab83b3"} Nov 29 04:57:24 crc kubenswrapper[4799]: I1129 04:57:24.710631 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a02375c1c673ae40e9efcab03118a1d57fe651613ed3023167240d74aab83b3" Nov 29 04:57:24 crc kubenswrapper[4799]: I1129 04:57:24.713475 4799 generic.go:334] "Generic (PLEG): container finished" podID="d4741f22-658c-40af-bba0-b36c491be636" containerID="cb2cd9b60acdc3381d394780aa0fc5cf6c8d8b1a94029038393ac04292fbfb98" exitCode=0 Nov 29 04:57:24 crc kubenswrapper[4799]: I1129 04:57:24.713527 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d4741f22-658c-40af-bba0-b36c491be636","Type":"ContainerDied","Data":"cb2cd9b60acdc3381d394780aa0fc5cf6c8d8b1a94029038393ac04292fbfb98"} Nov 29 04:57:24 crc kubenswrapper[4799]: I1129 04:57:24.715876 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"66c61a7f-a655-46e9-8a85-de7ad52bfe6f","Type":"ContainerStarted","Data":"ea8d268614151195fe9b4aa3ddd481ca2dc5e8091589f5a2e32645107358a927"} Nov 29 04:57:24 crc kubenswrapper[4799]: I1129 04:57:24.715986 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"66c61a7f-a655-46e9-8a85-de7ad52bfe6f","Type":"ContainerStarted","Data":"33bc1604771310b97be652234b9b2497cdbcc0d8505ab90cd91f4bb7205e2a72"} Nov 29 04:57:24 crc kubenswrapper[4799]: I1129 04:57:24.741638 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.3101109380000002 podStartE2EDuration="2.741612336s" podCreationTimestamp="2025-11-29 04:57:22 +0000 UTC" firstStartedPulling="2025-11-29 04:57:23.663778087 +0000 UTC m=+1119.306708477" lastFinishedPulling="2025-11-29 04:57:24.095279475 +0000 UTC m=+1119.738209875" observedRunningTime="2025-11-29 04:57:24.73550066 +0000 UTC m=+1120.378431080" watchObservedRunningTime="2025-11-29 04:57:24.741612336 +0000 UTC m=+1120.384542736" Nov 29 04:57:24 crc kubenswrapper[4799]: I1129 04:57:24.858942 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 29 04:57:24 crc kubenswrapper[4799]: I1129 04:57:24.859282 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ae86ab9e-2d28-4074-8364-9d1a01d77b62" containerName="nova-api-log" containerID="cri-o://c1d81c8095f715598cc9fb838de2ce04ef7ca5c82f6b28b2c88ab54e618a9fea" gracePeriod=30 Nov 29 04:57:24 crc kubenswrapper[4799]: I1129 04:57:24.859344 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ae86ab9e-2d28-4074-8364-9d1a01d77b62" containerName="nova-api-api" containerID="cri-o://ac913cd6f6b3c6cd2d97df30c634ac594219e8c013eb65754ed899e61dbaf202" gracePeriod=30 Nov 29 04:57:24 crc kubenswrapper[4799]: I1129 04:57:24.964495 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 29 04:57:25 crc kubenswrapper[4799]: I1129 04:57:25.190263 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-s5l5s" Nov 29 04:57:25 crc kubenswrapper[4799]: I1129 04:57:25.336627 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-67b47\" (UniqueName: \"kubernetes.io/projected/302ec8e8-d6a9-4bf2-a586-d6e95fcb232a-kube-api-access-67b47\") pod \"302ec8e8-d6a9-4bf2-a586-d6e95fcb232a\" (UID: \"302ec8e8-d6a9-4bf2-a586-d6e95fcb232a\") " Nov 29 04:57:25 crc kubenswrapper[4799]: I1129 04:57:25.336759 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/302ec8e8-d6a9-4bf2-a586-d6e95fcb232a-combined-ca-bundle\") pod \"302ec8e8-d6a9-4bf2-a586-d6e95fcb232a\" (UID: \"302ec8e8-d6a9-4bf2-a586-d6e95fcb232a\") " Nov 29 04:57:25 crc kubenswrapper[4799]: I1129 04:57:25.336906 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/302ec8e8-d6a9-4bf2-a586-d6e95fcb232a-scripts\") pod \"302ec8e8-d6a9-4bf2-a586-d6e95fcb232a\" (UID: \"302ec8e8-d6a9-4bf2-a586-d6e95fcb232a\") " Nov 29 04:57:25 crc kubenswrapper[4799]: I1129 04:57:25.336969 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/302ec8e8-d6a9-4bf2-a586-d6e95fcb232a-config-data\") pod \"302ec8e8-d6a9-4bf2-a586-d6e95fcb232a\" (UID: \"302ec8e8-d6a9-4bf2-a586-d6e95fcb232a\") " Nov 29 04:57:25 crc kubenswrapper[4799]: I1129 04:57:25.342566 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/302ec8e8-d6a9-4bf2-a586-d6e95fcb232a-scripts" (OuterVolumeSpecName: "scripts") pod "302ec8e8-d6a9-4bf2-a586-d6e95fcb232a" (UID: "302ec8e8-d6a9-4bf2-a586-d6e95fcb232a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:57:25 crc kubenswrapper[4799]: I1129 04:57:25.343817 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/302ec8e8-d6a9-4bf2-a586-d6e95fcb232a-kube-api-access-67b47" (OuterVolumeSpecName: "kube-api-access-67b47") pod "302ec8e8-d6a9-4bf2-a586-d6e95fcb232a" (UID: "302ec8e8-d6a9-4bf2-a586-d6e95fcb232a"). InnerVolumeSpecName "kube-api-access-67b47". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:57:25 crc kubenswrapper[4799]: I1129 04:57:25.369955 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/302ec8e8-d6a9-4bf2-a586-d6e95fcb232a-config-data" (OuterVolumeSpecName: "config-data") pod "302ec8e8-d6a9-4bf2-a586-d6e95fcb232a" (UID: "302ec8e8-d6a9-4bf2-a586-d6e95fcb232a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:57:25 crc kubenswrapper[4799]: I1129 04:57:25.370021 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/302ec8e8-d6a9-4bf2-a586-d6e95fcb232a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "302ec8e8-d6a9-4bf2-a586-d6e95fcb232a" (UID: "302ec8e8-d6a9-4bf2-a586-d6e95fcb232a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:57:25 crc kubenswrapper[4799]: I1129 04:57:25.439001 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/302ec8e8-d6a9-4bf2-a586-d6e95fcb232a-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:25 crc kubenswrapper[4799]: I1129 04:57:25.439044 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/302ec8e8-d6a9-4bf2-a586-d6e95fcb232a-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:25 crc kubenswrapper[4799]: I1129 04:57:25.439060 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-67b47\" (UniqueName: \"kubernetes.io/projected/302ec8e8-d6a9-4bf2-a586-d6e95fcb232a-kube-api-access-67b47\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:25 crc kubenswrapper[4799]: I1129 04:57:25.439074 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/302ec8e8-d6a9-4bf2-a586-d6e95fcb232a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:25 crc kubenswrapper[4799]: I1129 04:57:25.735847 4799 generic.go:334] "Generic (PLEG): container finished" podID="ae86ab9e-2d28-4074-8364-9d1a01d77b62" containerID="c1d81c8095f715598cc9fb838de2ce04ef7ca5c82f6b28b2c88ab54e618a9fea" exitCode=143 Nov 29 04:57:25 crc kubenswrapper[4799]: I1129 04:57:25.735959 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ae86ab9e-2d28-4074-8364-9d1a01d77b62","Type":"ContainerDied","Data":"c1d81c8095f715598cc9fb838de2ce04ef7ca5c82f6b28b2c88ab54e618a9fea"} Nov 29 04:57:25 crc kubenswrapper[4799]: I1129 04:57:25.753517 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-s5l5s" Nov 29 04:57:25 crc kubenswrapper[4799]: I1129 04:57:25.756780 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-s5l5s" event={"ID":"302ec8e8-d6a9-4bf2-a586-d6e95fcb232a","Type":"ContainerDied","Data":"4bb6de749e7dcbf61ba7e6655870209b24923730cbc44cce6cf6b6ed2b0ef388"} Nov 29 04:57:25 crc kubenswrapper[4799]: I1129 04:57:25.756846 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4bb6de749e7dcbf61ba7e6655870209b24923730cbc44cce6cf6b6ed2b0ef388" Nov 29 04:57:25 crc kubenswrapper[4799]: I1129 04:57:25.756870 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 29 04:57:25 crc kubenswrapper[4799]: I1129 04:57:25.772919 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 29 04:57:25 crc kubenswrapper[4799]: E1129 04:57:25.773496 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a28f76a7-54e5-4434-ba25-d3d046b30d5a" containerName="nova-manage" Nov 29 04:57:25 crc kubenswrapper[4799]: I1129 04:57:25.773516 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a28f76a7-54e5-4434-ba25-d3d046b30d5a" containerName="nova-manage" Nov 29 04:57:25 crc kubenswrapper[4799]: E1129 04:57:25.773535 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="302ec8e8-d6a9-4bf2-a586-d6e95fcb232a" containerName="nova-cell1-conductor-db-sync" Nov 29 04:57:25 crc kubenswrapper[4799]: I1129 04:57:25.773548 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="302ec8e8-d6a9-4bf2-a586-d6e95fcb232a" containerName="nova-cell1-conductor-db-sync" Nov 29 04:57:25 crc kubenswrapper[4799]: I1129 04:57:25.773802 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="302ec8e8-d6a9-4bf2-a586-d6e95fcb232a" containerName="nova-cell1-conductor-db-sync" Nov 29 04:57:25 crc kubenswrapper[4799]: I1129 04:57:25.773835 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="a28f76a7-54e5-4434-ba25-d3d046b30d5a" containerName="nova-manage" Nov 29 04:57:25 crc kubenswrapper[4799]: I1129 04:57:25.774674 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 29 04:57:25 crc kubenswrapper[4799]: I1129 04:57:25.780601 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 29 04:57:25 crc kubenswrapper[4799]: I1129 04:57:25.793281 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 29 04:57:25 crc kubenswrapper[4799]: I1129 04:57:25.849357 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27b8e913-bbd1-4a53-ab29-9b31ab8c64c8-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"27b8e913-bbd1-4a53-ab29-9b31ab8c64c8\") " pod="openstack/nova-cell1-conductor-0" Nov 29 04:57:25 crc kubenswrapper[4799]: I1129 04:57:25.849544 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4f9qn\" (UniqueName: \"kubernetes.io/projected/27b8e913-bbd1-4a53-ab29-9b31ab8c64c8-kube-api-access-4f9qn\") pod \"nova-cell1-conductor-0\" (UID: \"27b8e913-bbd1-4a53-ab29-9b31ab8c64c8\") " pod="openstack/nova-cell1-conductor-0" Nov 29 04:57:25 crc kubenswrapper[4799]: I1129 04:57:25.849653 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27b8e913-bbd1-4a53-ab29-9b31ab8c64c8-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"27b8e913-bbd1-4a53-ab29-9b31ab8c64c8\") " pod="openstack/nova-cell1-conductor-0" Nov 29 04:57:25 crc kubenswrapper[4799]: I1129 04:57:25.951319 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4f9qn\" (UniqueName: \"kubernetes.io/projected/27b8e913-bbd1-4a53-ab29-9b31ab8c64c8-kube-api-access-4f9qn\") pod \"nova-cell1-conductor-0\" (UID: \"27b8e913-bbd1-4a53-ab29-9b31ab8c64c8\") " pod="openstack/nova-cell1-conductor-0" Nov 29 04:57:25 crc kubenswrapper[4799]: I1129 04:57:25.951420 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27b8e913-bbd1-4a53-ab29-9b31ab8c64c8-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"27b8e913-bbd1-4a53-ab29-9b31ab8c64c8\") " pod="openstack/nova-cell1-conductor-0" Nov 29 04:57:25 crc kubenswrapper[4799]: I1129 04:57:25.951523 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27b8e913-bbd1-4a53-ab29-9b31ab8c64c8-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"27b8e913-bbd1-4a53-ab29-9b31ab8c64c8\") " pod="openstack/nova-cell1-conductor-0" Nov 29 04:57:25 crc kubenswrapper[4799]: I1129 04:57:25.956389 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27b8e913-bbd1-4a53-ab29-9b31ab8c64c8-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"27b8e913-bbd1-4a53-ab29-9b31ab8c64c8\") " pod="openstack/nova-cell1-conductor-0" Nov 29 04:57:25 crc kubenswrapper[4799]: I1129 04:57:25.958361 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27b8e913-bbd1-4a53-ab29-9b31ab8c64c8-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"27b8e913-bbd1-4a53-ab29-9b31ab8c64c8\") " pod="openstack/nova-cell1-conductor-0" Nov 29 04:57:25 crc kubenswrapper[4799]: I1129 04:57:25.977537 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4f9qn\" (UniqueName: \"kubernetes.io/projected/27b8e913-bbd1-4a53-ab29-9b31ab8c64c8-kube-api-access-4f9qn\") pod \"nova-cell1-conductor-0\" (UID: \"27b8e913-bbd1-4a53-ab29-9b31ab8c64c8\") " pod="openstack/nova-cell1-conductor-0" Nov 29 04:57:26 crc kubenswrapper[4799]: I1129 04:57:26.112820 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 29 04:57:26 crc kubenswrapper[4799]: I1129 04:57:26.600750 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 29 04:57:26 crc kubenswrapper[4799]: I1129 04:57:26.766655 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"27b8e913-bbd1-4a53-ab29-9b31ab8c64c8","Type":"ContainerStarted","Data":"f1c8b45a02bb6869b7f94952dac9c5f2b201a825d22b502ee336a3d39f06e076"} Nov 29 04:57:26 crc kubenswrapper[4799]: I1129 04:57:26.767026 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="06f32a7b-0ded-4927-9b8b-55fc1ecbb296" containerName="nova-scheduler-scheduler" containerID="cri-o://eb674ecc4ca4399e936f237d941a66c8354d24c9fb184e868257e4e51dca7f89" gracePeriod=30 Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.251950 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.395368 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4741f22-658c-40af-bba0-b36c491be636-combined-ca-bundle\") pod \"d4741f22-658c-40af-bba0-b36c491be636\" (UID: \"d4741f22-658c-40af-bba0-b36c491be636\") " Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.395558 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d4741f22-658c-40af-bba0-b36c491be636-run-httpd\") pod \"d4741f22-658c-40af-bba0-b36c491be636\" (UID: \"d4741f22-658c-40af-bba0-b36c491be636\") " Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.395587 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4741f22-658c-40af-bba0-b36c491be636-config-data\") pod \"d4741f22-658c-40af-bba0-b36c491be636\" (UID: \"d4741f22-658c-40af-bba0-b36c491be636\") " Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.395658 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d4741f22-658c-40af-bba0-b36c491be636-sg-core-conf-yaml\") pod \"d4741f22-658c-40af-bba0-b36c491be636\" (UID: \"d4741f22-658c-40af-bba0-b36c491be636\") " Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.395739 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4741f22-658c-40af-bba0-b36c491be636-scripts\") pod \"d4741f22-658c-40af-bba0-b36c491be636\" (UID: \"d4741f22-658c-40af-bba0-b36c491be636\") " Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.395779 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qht5x\" (UniqueName: \"kubernetes.io/projected/d4741f22-658c-40af-bba0-b36c491be636-kube-api-access-qht5x\") pod \"d4741f22-658c-40af-bba0-b36c491be636\" (UID: \"d4741f22-658c-40af-bba0-b36c491be636\") " Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.395848 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d4741f22-658c-40af-bba0-b36c491be636-log-httpd\") pod \"d4741f22-658c-40af-bba0-b36c491be636\" (UID: \"d4741f22-658c-40af-bba0-b36c491be636\") " Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.396058 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4741f22-658c-40af-bba0-b36c491be636-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d4741f22-658c-40af-bba0-b36c491be636" (UID: "d4741f22-658c-40af-bba0-b36c491be636"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.396518 4799 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d4741f22-658c-40af-bba0-b36c491be636-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.396628 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4741f22-658c-40af-bba0-b36c491be636-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d4741f22-658c-40af-bba0-b36c491be636" (UID: "d4741f22-658c-40af-bba0-b36c491be636"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.403968 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4741f22-658c-40af-bba0-b36c491be636-kube-api-access-qht5x" (OuterVolumeSpecName: "kube-api-access-qht5x") pod "d4741f22-658c-40af-bba0-b36c491be636" (UID: "d4741f22-658c-40af-bba0-b36c491be636"). InnerVolumeSpecName "kube-api-access-qht5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.421946 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4741f22-658c-40af-bba0-b36c491be636-scripts" (OuterVolumeSpecName: "scripts") pod "d4741f22-658c-40af-bba0-b36c491be636" (UID: "d4741f22-658c-40af-bba0-b36c491be636"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.425928 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4741f22-658c-40af-bba0-b36c491be636-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d4741f22-658c-40af-bba0-b36c491be636" (UID: "d4741f22-658c-40af-bba0-b36c491be636"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.477509 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4741f22-658c-40af-bba0-b36c491be636-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d4741f22-658c-40af-bba0-b36c491be636" (UID: "d4741f22-658c-40af-bba0-b36c491be636"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.498170 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4741f22-658c-40af-bba0-b36c491be636-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.498424 4799 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d4741f22-658c-40af-bba0-b36c491be636-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.498509 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4741f22-658c-40af-bba0-b36c491be636-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.498585 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qht5x\" (UniqueName: \"kubernetes.io/projected/d4741f22-658c-40af-bba0-b36c491be636-kube-api-access-qht5x\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.498663 4799 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d4741f22-658c-40af-bba0-b36c491be636-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.504541 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4741f22-658c-40af-bba0-b36c491be636-config-data" (OuterVolumeSpecName: "config-data") pod "d4741f22-658c-40af-bba0-b36c491be636" (UID: "d4741f22-658c-40af-bba0-b36c491be636"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.600944 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4741f22-658c-40af-bba0-b36c491be636-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.777635 4799 generic.go:334] "Generic (PLEG): container finished" podID="d4741f22-658c-40af-bba0-b36c491be636" containerID="ff92573600da0847b820dabc90979d71078d06a62fe0d43ca46e4ca0b36f4328" exitCode=0 Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.777720 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d4741f22-658c-40af-bba0-b36c491be636","Type":"ContainerDied","Data":"ff92573600da0847b820dabc90979d71078d06a62fe0d43ca46e4ca0b36f4328"} Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.777865 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d4741f22-658c-40af-bba0-b36c491be636","Type":"ContainerDied","Data":"30734b3a95d85ea21cbe9e0d1f2c140cb31d17f67cf4b2cf8e610d4be32d771e"} Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.777897 4799 scope.go:117] "RemoveContainer" containerID="d67f65d1c9b133f52aae7c40e3e167681a53206dbdf3f6cf9575afc01f64f4b6" Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.778071 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.782378 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"27b8e913-bbd1-4a53-ab29-9b31ab8c64c8","Type":"ContainerStarted","Data":"dac56ce88701e8fdf38eb926d656d908f56c04192dc23494c7741da38d7a590f"} Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.782758 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.802568 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.802537437 podStartE2EDuration="2.802537437s" podCreationTimestamp="2025-11-29 04:57:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:57:27.80010255 +0000 UTC m=+1123.443032950" watchObservedRunningTime="2025-11-29 04:57:27.802537437 +0000 UTC m=+1123.445467837" Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.805394 4799 scope.go:117] "RemoveContainer" containerID="bc58c8ff01459001f39e28a7979a9c4bfd0e961f421845b23dd2cff71163ceb3" Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.827316 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.835696 4799 scope.go:117] "RemoveContainer" containerID="ff92573600da0847b820dabc90979d71078d06a62fe0d43ca46e4ca0b36f4328" Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.842443 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.867272 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 29 04:57:27 crc kubenswrapper[4799]: E1129 04:57:27.867774 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4741f22-658c-40af-bba0-b36c491be636" containerName="sg-core" Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.867814 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4741f22-658c-40af-bba0-b36c491be636" containerName="sg-core" Nov 29 04:57:27 crc kubenswrapper[4799]: E1129 04:57:27.867837 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4741f22-658c-40af-bba0-b36c491be636" containerName="ceilometer-central-agent" Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.867843 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4741f22-658c-40af-bba0-b36c491be636" containerName="ceilometer-central-agent" Nov 29 04:57:27 crc kubenswrapper[4799]: E1129 04:57:27.867851 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4741f22-658c-40af-bba0-b36c491be636" containerName="proxy-httpd" Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.867857 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4741f22-658c-40af-bba0-b36c491be636" containerName="proxy-httpd" Nov 29 04:57:27 crc kubenswrapper[4799]: E1129 04:57:27.867868 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4741f22-658c-40af-bba0-b36c491be636" containerName="ceilometer-notification-agent" Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.867873 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4741f22-658c-40af-bba0-b36c491be636" containerName="ceilometer-notification-agent" Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.868292 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4741f22-658c-40af-bba0-b36c491be636" containerName="sg-core" Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.868320 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4741f22-658c-40af-bba0-b36c491be636" containerName="ceilometer-notification-agent" Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.868330 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4741f22-658c-40af-bba0-b36c491be636" containerName="proxy-httpd" Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.868343 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4741f22-658c-40af-bba0-b36c491be636" containerName="ceilometer-central-agent" Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.871475 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.873764 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.874028 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.875668 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.875952 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.884701 4799 scope.go:117] "RemoveContainer" containerID="cb2cd9b60acdc3381d394780aa0fc5cf6c8d8b1a94029038393ac04292fbfb98" Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.918019 4799 scope.go:117] "RemoveContainer" containerID="d67f65d1c9b133f52aae7c40e3e167681a53206dbdf3f6cf9575afc01f64f4b6" Nov 29 04:57:27 crc kubenswrapper[4799]: E1129 04:57:27.918578 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d67f65d1c9b133f52aae7c40e3e167681a53206dbdf3f6cf9575afc01f64f4b6\": container with ID starting with d67f65d1c9b133f52aae7c40e3e167681a53206dbdf3f6cf9575afc01f64f4b6 not found: ID does not exist" containerID="d67f65d1c9b133f52aae7c40e3e167681a53206dbdf3f6cf9575afc01f64f4b6" Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.918673 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d67f65d1c9b133f52aae7c40e3e167681a53206dbdf3f6cf9575afc01f64f4b6"} err="failed to get container status \"d67f65d1c9b133f52aae7c40e3e167681a53206dbdf3f6cf9575afc01f64f4b6\": rpc error: code = NotFound desc = could not find container \"d67f65d1c9b133f52aae7c40e3e167681a53206dbdf3f6cf9575afc01f64f4b6\": container with ID starting with d67f65d1c9b133f52aae7c40e3e167681a53206dbdf3f6cf9575afc01f64f4b6 not found: ID does not exist" Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.918751 4799 scope.go:117] "RemoveContainer" containerID="bc58c8ff01459001f39e28a7979a9c4bfd0e961f421845b23dd2cff71163ceb3" Nov 29 04:57:27 crc kubenswrapper[4799]: E1129 04:57:27.919222 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc58c8ff01459001f39e28a7979a9c4bfd0e961f421845b23dd2cff71163ceb3\": container with ID starting with bc58c8ff01459001f39e28a7979a9c4bfd0e961f421845b23dd2cff71163ceb3 not found: ID does not exist" containerID="bc58c8ff01459001f39e28a7979a9c4bfd0e961f421845b23dd2cff71163ceb3" Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.919296 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc58c8ff01459001f39e28a7979a9c4bfd0e961f421845b23dd2cff71163ceb3"} err="failed to get container status \"bc58c8ff01459001f39e28a7979a9c4bfd0e961f421845b23dd2cff71163ceb3\": rpc error: code = NotFound desc = could not find container \"bc58c8ff01459001f39e28a7979a9c4bfd0e961f421845b23dd2cff71163ceb3\": container with ID starting with bc58c8ff01459001f39e28a7979a9c4bfd0e961f421845b23dd2cff71163ceb3 not found: ID does not exist" Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.919367 4799 scope.go:117] "RemoveContainer" containerID="ff92573600da0847b820dabc90979d71078d06a62fe0d43ca46e4ca0b36f4328" Nov 29 04:57:27 crc kubenswrapper[4799]: E1129 04:57:27.919618 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff92573600da0847b820dabc90979d71078d06a62fe0d43ca46e4ca0b36f4328\": container with ID starting with ff92573600da0847b820dabc90979d71078d06a62fe0d43ca46e4ca0b36f4328 not found: ID does not exist" containerID="ff92573600da0847b820dabc90979d71078d06a62fe0d43ca46e4ca0b36f4328" Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.919692 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff92573600da0847b820dabc90979d71078d06a62fe0d43ca46e4ca0b36f4328"} err="failed to get container status \"ff92573600da0847b820dabc90979d71078d06a62fe0d43ca46e4ca0b36f4328\": rpc error: code = NotFound desc = could not find container \"ff92573600da0847b820dabc90979d71078d06a62fe0d43ca46e4ca0b36f4328\": container with ID starting with ff92573600da0847b820dabc90979d71078d06a62fe0d43ca46e4ca0b36f4328 not found: ID does not exist" Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.919752 4799 scope.go:117] "RemoveContainer" containerID="cb2cd9b60acdc3381d394780aa0fc5cf6c8d8b1a94029038393ac04292fbfb98" Nov 29 04:57:27 crc kubenswrapper[4799]: E1129 04:57:27.920209 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb2cd9b60acdc3381d394780aa0fc5cf6c8d8b1a94029038393ac04292fbfb98\": container with ID starting with cb2cd9b60acdc3381d394780aa0fc5cf6c8d8b1a94029038393ac04292fbfb98 not found: ID does not exist" containerID="cb2cd9b60acdc3381d394780aa0fc5cf6c8d8b1a94029038393ac04292fbfb98" Nov 29 04:57:27 crc kubenswrapper[4799]: I1129 04:57:27.920290 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb2cd9b60acdc3381d394780aa0fc5cf6c8d8b1a94029038393ac04292fbfb98"} err="failed to get container status \"cb2cd9b60acdc3381d394780aa0fc5cf6c8d8b1a94029038393ac04292fbfb98\": rpc error: code = NotFound desc = could not find container \"cb2cd9b60acdc3381d394780aa0fc5cf6c8d8b1a94029038393ac04292fbfb98\": container with ID starting with cb2cd9b60acdc3381d394780aa0fc5cf6c8d8b1a94029038393ac04292fbfb98 not found: ID does not exist" Nov 29 04:57:28 crc kubenswrapper[4799]: I1129 04:57:28.009384 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33323158-ffcc-4f35-9797-e1fb18c3f0f1-run-httpd\") pod \"ceilometer-0\" (UID: \"33323158-ffcc-4f35-9797-e1fb18c3f0f1\") " pod="openstack/ceilometer-0" Nov 29 04:57:28 crc kubenswrapper[4799]: I1129 04:57:28.009636 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/33323158-ffcc-4f35-9797-e1fb18c3f0f1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"33323158-ffcc-4f35-9797-e1fb18c3f0f1\") " pod="openstack/ceilometer-0" Nov 29 04:57:28 crc kubenswrapper[4799]: I1129 04:57:28.009778 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33323158-ffcc-4f35-9797-e1fb18c3f0f1-scripts\") pod \"ceilometer-0\" (UID: \"33323158-ffcc-4f35-9797-e1fb18c3f0f1\") " pod="openstack/ceilometer-0" Nov 29 04:57:28 crc kubenswrapper[4799]: I1129 04:57:28.009990 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zghtt\" (UniqueName: \"kubernetes.io/projected/33323158-ffcc-4f35-9797-e1fb18c3f0f1-kube-api-access-zghtt\") pod \"ceilometer-0\" (UID: \"33323158-ffcc-4f35-9797-e1fb18c3f0f1\") " pod="openstack/ceilometer-0" Nov 29 04:57:28 crc kubenswrapper[4799]: I1129 04:57:28.010096 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33323158-ffcc-4f35-9797-e1fb18c3f0f1-config-data\") pod \"ceilometer-0\" (UID: \"33323158-ffcc-4f35-9797-e1fb18c3f0f1\") " pod="openstack/ceilometer-0" Nov 29 04:57:28 crc kubenswrapper[4799]: I1129 04:57:28.010180 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/33323158-ffcc-4f35-9797-e1fb18c3f0f1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"33323158-ffcc-4f35-9797-e1fb18c3f0f1\") " pod="openstack/ceilometer-0" Nov 29 04:57:28 crc kubenswrapper[4799]: I1129 04:57:28.010333 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33323158-ffcc-4f35-9797-e1fb18c3f0f1-log-httpd\") pod \"ceilometer-0\" (UID: \"33323158-ffcc-4f35-9797-e1fb18c3f0f1\") " pod="openstack/ceilometer-0" Nov 29 04:57:28 crc kubenswrapper[4799]: I1129 04:57:28.010400 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33323158-ffcc-4f35-9797-e1fb18c3f0f1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"33323158-ffcc-4f35-9797-e1fb18c3f0f1\") " pod="openstack/ceilometer-0" Nov 29 04:57:28 crc kubenswrapper[4799]: E1129 04:57:28.012631 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="eb674ecc4ca4399e936f237d941a66c8354d24c9fb184e868257e4e51dca7f89" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 29 04:57:28 crc kubenswrapper[4799]: E1129 04:57:28.013937 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="eb674ecc4ca4399e936f237d941a66c8354d24c9fb184e868257e4e51dca7f89" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 29 04:57:28 crc kubenswrapper[4799]: E1129 04:57:28.015502 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="eb674ecc4ca4399e936f237d941a66c8354d24c9fb184e868257e4e51dca7f89" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 29 04:57:28 crc kubenswrapper[4799]: E1129 04:57:28.015685 4799 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="06f32a7b-0ded-4927-9b8b-55fc1ecbb296" containerName="nova-scheduler-scheduler" Nov 29 04:57:28 crc kubenswrapper[4799]: I1129 04:57:28.112366 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33323158-ffcc-4f35-9797-e1fb18c3f0f1-log-httpd\") pod \"ceilometer-0\" (UID: \"33323158-ffcc-4f35-9797-e1fb18c3f0f1\") " pod="openstack/ceilometer-0" Nov 29 04:57:28 crc kubenswrapper[4799]: I1129 04:57:28.113221 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33323158-ffcc-4f35-9797-e1fb18c3f0f1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"33323158-ffcc-4f35-9797-e1fb18c3f0f1\") " pod="openstack/ceilometer-0" Nov 29 04:57:28 crc kubenswrapper[4799]: I1129 04:57:28.113184 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33323158-ffcc-4f35-9797-e1fb18c3f0f1-log-httpd\") pod \"ceilometer-0\" (UID: \"33323158-ffcc-4f35-9797-e1fb18c3f0f1\") " pod="openstack/ceilometer-0" Nov 29 04:57:28 crc kubenswrapper[4799]: I1129 04:57:28.114286 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33323158-ffcc-4f35-9797-e1fb18c3f0f1-run-httpd\") pod \"ceilometer-0\" (UID: \"33323158-ffcc-4f35-9797-e1fb18c3f0f1\") " pod="openstack/ceilometer-0" Nov 29 04:57:28 crc kubenswrapper[4799]: I1129 04:57:28.114394 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/33323158-ffcc-4f35-9797-e1fb18c3f0f1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"33323158-ffcc-4f35-9797-e1fb18c3f0f1\") " pod="openstack/ceilometer-0" Nov 29 04:57:28 crc kubenswrapper[4799]: I1129 04:57:28.114564 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33323158-ffcc-4f35-9797-e1fb18c3f0f1-scripts\") pod \"ceilometer-0\" (UID: \"33323158-ffcc-4f35-9797-e1fb18c3f0f1\") " pod="openstack/ceilometer-0" Nov 29 04:57:28 crc kubenswrapper[4799]: I1129 04:57:28.114720 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zghtt\" (UniqueName: \"kubernetes.io/projected/33323158-ffcc-4f35-9797-e1fb18c3f0f1-kube-api-access-zghtt\") pod \"ceilometer-0\" (UID: \"33323158-ffcc-4f35-9797-e1fb18c3f0f1\") " pod="openstack/ceilometer-0" Nov 29 04:57:28 crc kubenswrapper[4799]: I1129 04:57:28.115897 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33323158-ffcc-4f35-9797-e1fb18c3f0f1-run-httpd\") pod \"ceilometer-0\" (UID: \"33323158-ffcc-4f35-9797-e1fb18c3f0f1\") " pod="openstack/ceilometer-0" Nov 29 04:57:28 crc kubenswrapper[4799]: I1129 04:57:28.121165 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33323158-ffcc-4f35-9797-e1fb18c3f0f1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"33323158-ffcc-4f35-9797-e1fb18c3f0f1\") " pod="openstack/ceilometer-0" Nov 29 04:57:28 crc kubenswrapper[4799]: I1129 04:57:28.120070 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33323158-ffcc-4f35-9797-e1fb18c3f0f1-config-data\") pod \"ceilometer-0\" (UID: \"33323158-ffcc-4f35-9797-e1fb18c3f0f1\") " pod="openstack/ceilometer-0" Nov 29 04:57:28 crc kubenswrapper[4799]: I1129 04:57:28.121424 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/33323158-ffcc-4f35-9797-e1fb18c3f0f1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"33323158-ffcc-4f35-9797-e1fb18c3f0f1\") " pod="openstack/ceilometer-0" Nov 29 04:57:28 crc kubenswrapper[4799]: I1129 04:57:28.121871 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33323158-ffcc-4f35-9797-e1fb18c3f0f1-scripts\") pod \"ceilometer-0\" (UID: \"33323158-ffcc-4f35-9797-e1fb18c3f0f1\") " pod="openstack/ceilometer-0" Nov 29 04:57:28 crc kubenswrapper[4799]: I1129 04:57:28.123648 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/33323158-ffcc-4f35-9797-e1fb18c3f0f1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"33323158-ffcc-4f35-9797-e1fb18c3f0f1\") " pod="openstack/ceilometer-0" Nov 29 04:57:28 crc kubenswrapper[4799]: I1129 04:57:28.125621 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/33323158-ffcc-4f35-9797-e1fb18c3f0f1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"33323158-ffcc-4f35-9797-e1fb18c3f0f1\") " pod="openstack/ceilometer-0" Nov 29 04:57:28 crc kubenswrapper[4799]: I1129 04:57:28.132477 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33323158-ffcc-4f35-9797-e1fb18c3f0f1-config-data\") pod \"ceilometer-0\" (UID: \"33323158-ffcc-4f35-9797-e1fb18c3f0f1\") " pod="openstack/ceilometer-0" Nov 29 04:57:28 crc kubenswrapper[4799]: I1129 04:57:28.145238 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zghtt\" (UniqueName: \"kubernetes.io/projected/33323158-ffcc-4f35-9797-e1fb18c3f0f1-kube-api-access-zghtt\") pod \"ceilometer-0\" (UID: \"33323158-ffcc-4f35-9797-e1fb18c3f0f1\") " pod="openstack/ceilometer-0" Nov 29 04:57:28 crc kubenswrapper[4799]: I1129 04:57:28.200701 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 04:57:28 crc kubenswrapper[4799]: I1129 04:57:28.305012 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-566b5b7845-8frfb" Nov 29 04:57:28 crc kubenswrapper[4799]: I1129 04:57:28.385022 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d97fcdd8f-9dgkt"] Nov 29 04:57:28 crc kubenswrapper[4799]: I1129 04:57:28.385580 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6d97fcdd8f-9dgkt" podUID="c419e681-9099-4422-b04e-8faccbedc127" containerName="dnsmasq-dns" containerID="cri-o://296552b3cd8ba3aa9aef20b4bc6845e18cb64ff6248d63a812dda6adcbf9457f" gracePeriod=10 Nov 29 04:57:28 crc kubenswrapper[4799]: I1129 04:57:28.670566 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4741f22-658c-40af-bba0-b36c491be636" path="/var/lib/kubelet/pods/d4741f22-658c-40af-bba0-b36c491be636/volumes" Nov 29 04:57:28 crc kubenswrapper[4799]: I1129 04:57:28.733444 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 29 04:57:28 crc kubenswrapper[4799]: W1129 04:57:28.753331 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod33323158_ffcc_4f35_9797_e1fb18c3f0f1.slice/crio-0a7339bd71d1f8c83aaef56fd7f9e1c198ef00141545d1a6859fd487919b6ffb WatchSource:0}: Error finding container 0a7339bd71d1f8c83aaef56fd7f9e1c198ef00141545d1a6859fd487919b6ffb: Status 404 returned error can't find the container with id 0a7339bd71d1f8c83aaef56fd7f9e1c198ef00141545d1a6859fd487919b6ffb Nov 29 04:57:28 crc kubenswrapper[4799]: I1129 04:57:28.801028 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33323158-ffcc-4f35-9797-e1fb18c3f0f1","Type":"ContainerStarted","Data":"0a7339bd71d1f8c83aaef56fd7f9e1c198ef00141545d1a6859fd487919b6ffb"} Nov 29 04:57:28 crc kubenswrapper[4799]: I1129 04:57:28.808656 4799 generic.go:334] "Generic (PLEG): container finished" podID="c419e681-9099-4422-b04e-8faccbedc127" containerID="296552b3cd8ba3aa9aef20b4bc6845e18cb64ff6248d63a812dda6adcbf9457f" exitCode=0 Nov 29 04:57:28 crc kubenswrapper[4799]: I1129 04:57:28.808746 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d97fcdd8f-9dgkt" event={"ID":"c419e681-9099-4422-b04e-8faccbedc127","Type":"ContainerDied","Data":"296552b3cd8ba3aa9aef20b4bc6845e18cb64ff6248d63a812dda6adcbf9457f"} Nov 29 04:57:28 crc kubenswrapper[4799]: I1129 04:57:28.825469 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d97fcdd8f-9dgkt" Nov 29 04:57:28 crc kubenswrapper[4799]: I1129 04:57:28.941499 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gzs8q\" (UniqueName: \"kubernetes.io/projected/c419e681-9099-4422-b04e-8faccbedc127-kube-api-access-gzs8q\") pod \"c419e681-9099-4422-b04e-8faccbedc127\" (UID: \"c419e681-9099-4422-b04e-8faccbedc127\") " Nov 29 04:57:28 crc kubenswrapper[4799]: I1129 04:57:28.941719 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c419e681-9099-4422-b04e-8faccbedc127-dns-svc\") pod \"c419e681-9099-4422-b04e-8faccbedc127\" (UID: \"c419e681-9099-4422-b04e-8faccbedc127\") " Nov 29 04:57:28 crc kubenswrapper[4799]: I1129 04:57:28.941807 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c419e681-9099-4422-b04e-8faccbedc127-config\") pod \"c419e681-9099-4422-b04e-8faccbedc127\" (UID: \"c419e681-9099-4422-b04e-8faccbedc127\") " Nov 29 04:57:28 crc kubenswrapper[4799]: I1129 04:57:28.941922 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c419e681-9099-4422-b04e-8faccbedc127-ovsdbserver-sb\") pod \"c419e681-9099-4422-b04e-8faccbedc127\" (UID: \"c419e681-9099-4422-b04e-8faccbedc127\") " Nov 29 04:57:28 crc kubenswrapper[4799]: I1129 04:57:28.941944 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c419e681-9099-4422-b04e-8faccbedc127-ovsdbserver-nb\") pod \"c419e681-9099-4422-b04e-8faccbedc127\" (UID: \"c419e681-9099-4422-b04e-8faccbedc127\") " Nov 29 04:57:28 crc kubenswrapper[4799]: I1129 04:57:28.949002 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c419e681-9099-4422-b04e-8faccbedc127-kube-api-access-gzs8q" (OuterVolumeSpecName: "kube-api-access-gzs8q") pod "c419e681-9099-4422-b04e-8faccbedc127" (UID: "c419e681-9099-4422-b04e-8faccbedc127"). InnerVolumeSpecName "kube-api-access-gzs8q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:57:28 crc kubenswrapper[4799]: I1129 04:57:28.999690 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c419e681-9099-4422-b04e-8faccbedc127-config" (OuterVolumeSpecName: "config") pod "c419e681-9099-4422-b04e-8faccbedc127" (UID: "c419e681-9099-4422-b04e-8faccbedc127"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:57:29 crc kubenswrapper[4799]: I1129 04:57:29.004735 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c419e681-9099-4422-b04e-8faccbedc127-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c419e681-9099-4422-b04e-8faccbedc127" (UID: "c419e681-9099-4422-b04e-8faccbedc127"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:57:29 crc kubenswrapper[4799]: I1129 04:57:29.012468 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c419e681-9099-4422-b04e-8faccbedc127-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c419e681-9099-4422-b04e-8faccbedc127" (UID: "c419e681-9099-4422-b04e-8faccbedc127"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:57:29 crc kubenswrapper[4799]: I1129 04:57:29.018888 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c419e681-9099-4422-b04e-8faccbedc127-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c419e681-9099-4422-b04e-8faccbedc127" (UID: "c419e681-9099-4422-b04e-8faccbedc127"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:57:29 crc kubenswrapper[4799]: I1129 04:57:29.045488 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c419e681-9099-4422-b04e-8faccbedc127-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:29 crc kubenswrapper[4799]: I1129 04:57:29.045532 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c419e681-9099-4422-b04e-8faccbedc127-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:29 crc kubenswrapper[4799]: I1129 04:57:29.045551 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gzs8q\" (UniqueName: \"kubernetes.io/projected/c419e681-9099-4422-b04e-8faccbedc127-kube-api-access-gzs8q\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:29 crc kubenswrapper[4799]: I1129 04:57:29.045572 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c419e681-9099-4422-b04e-8faccbedc127-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:29 crc kubenswrapper[4799]: I1129 04:57:29.045586 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c419e681-9099-4422-b04e-8faccbedc127-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:29 crc kubenswrapper[4799]: I1129 04:57:29.594733 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 29 04:57:29 crc kubenswrapper[4799]: I1129 04:57:29.658015 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae86ab9e-2d28-4074-8364-9d1a01d77b62-logs\") pod \"ae86ab9e-2d28-4074-8364-9d1a01d77b62\" (UID: \"ae86ab9e-2d28-4074-8364-9d1a01d77b62\") " Nov 29 04:57:29 crc kubenswrapper[4799]: I1129 04:57:29.658122 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae86ab9e-2d28-4074-8364-9d1a01d77b62-combined-ca-bundle\") pod \"ae86ab9e-2d28-4074-8364-9d1a01d77b62\" (UID: \"ae86ab9e-2d28-4074-8364-9d1a01d77b62\") " Nov 29 04:57:29 crc kubenswrapper[4799]: I1129 04:57:29.658182 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8v6rg\" (UniqueName: \"kubernetes.io/projected/ae86ab9e-2d28-4074-8364-9d1a01d77b62-kube-api-access-8v6rg\") pod \"ae86ab9e-2d28-4074-8364-9d1a01d77b62\" (UID: \"ae86ab9e-2d28-4074-8364-9d1a01d77b62\") " Nov 29 04:57:29 crc kubenswrapper[4799]: I1129 04:57:29.658404 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae86ab9e-2d28-4074-8364-9d1a01d77b62-config-data\") pod \"ae86ab9e-2d28-4074-8364-9d1a01d77b62\" (UID: \"ae86ab9e-2d28-4074-8364-9d1a01d77b62\") " Nov 29 04:57:29 crc kubenswrapper[4799]: I1129 04:57:29.658768 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae86ab9e-2d28-4074-8364-9d1a01d77b62-logs" (OuterVolumeSpecName: "logs") pod "ae86ab9e-2d28-4074-8364-9d1a01d77b62" (UID: "ae86ab9e-2d28-4074-8364-9d1a01d77b62"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:57:29 crc kubenswrapper[4799]: I1129 04:57:29.659092 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae86ab9e-2d28-4074-8364-9d1a01d77b62-logs\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:29 crc kubenswrapper[4799]: I1129 04:57:29.665328 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae86ab9e-2d28-4074-8364-9d1a01d77b62-kube-api-access-8v6rg" (OuterVolumeSpecName: "kube-api-access-8v6rg") pod "ae86ab9e-2d28-4074-8364-9d1a01d77b62" (UID: "ae86ab9e-2d28-4074-8364-9d1a01d77b62"). InnerVolumeSpecName "kube-api-access-8v6rg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:57:29 crc kubenswrapper[4799]: I1129 04:57:29.694851 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae86ab9e-2d28-4074-8364-9d1a01d77b62-config-data" (OuterVolumeSpecName: "config-data") pod "ae86ab9e-2d28-4074-8364-9d1a01d77b62" (UID: "ae86ab9e-2d28-4074-8364-9d1a01d77b62"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:57:29 crc kubenswrapper[4799]: I1129 04:57:29.708102 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae86ab9e-2d28-4074-8364-9d1a01d77b62-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ae86ab9e-2d28-4074-8364-9d1a01d77b62" (UID: "ae86ab9e-2d28-4074-8364-9d1a01d77b62"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:57:29 crc kubenswrapper[4799]: I1129 04:57:29.761376 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae86ab9e-2d28-4074-8364-9d1a01d77b62-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:29 crc kubenswrapper[4799]: I1129 04:57:29.761934 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae86ab9e-2d28-4074-8364-9d1a01d77b62-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:29 crc kubenswrapper[4799]: I1129 04:57:29.761989 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8v6rg\" (UniqueName: \"kubernetes.io/projected/ae86ab9e-2d28-4074-8364-9d1a01d77b62-kube-api-access-8v6rg\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:29 crc kubenswrapper[4799]: I1129 04:57:29.821066 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d97fcdd8f-9dgkt" event={"ID":"c419e681-9099-4422-b04e-8faccbedc127","Type":"ContainerDied","Data":"a503e1abc5f477580dbaaa3fab6e9bdca2000a937ddc7bb49c9988af8d365f93"} Nov 29 04:57:29 crc kubenswrapper[4799]: I1129 04:57:29.821158 4799 scope.go:117] "RemoveContainer" containerID="296552b3cd8ba3aa9aef20b4bc6845e18cb64ff6248d63a812dda6adcbf9457f" Nov 29 04:57:29 crc kubenswrapper[4799]: I1129 04:57:29.821924 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d97fcdd8f-9dgkt" Nov 29 04:57:29 crc kubenswrapper[4799]: I1129 04:57:29.822960 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33323158-ffcc-4f35-9797-e1fb18c3f0f1","Type":"ContainerStarted","Data":"9d0c07f09a3c0e24fb8bfd28d5f8c8ee44af9d249f5f1d1fb9b8f47a7f779370"} Nov 29 04:57:29 crc kubenswrapper[4799]: I1129 04:57:29.828275 4799 generic.go:334] "Generic (PLEG): container finished" podID="ae86ab9e-2d28-4074-8364-9d1a01d77b62" containerID="ac913cd6f6b3c6cd2d97df30c634ac594219e8c013eb65754ed899e61dbaf202" exitCode=0 Nov 29 04:57:29 crc kubenswrapper[4799]: I1129 04:57:29.828331 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ae86ab9e-2d28-4074-8364-9d1a01d77b62","Type":"ContainerDied","Data":"ac913cd6f6b3c6cd2d97df30c634ac594219e8c013eb65754ed899e61dbaf202"} Nov 29 04:57:29 crc kubenswrapper[4799]: I1129 04:57:29.828366 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ae86ab9e-2d28-4074-8364-9d1a01d77b62","Type":"ContainerDied","Data":"a9201a8933663eb523fc944cb23e7770003984dbf12f075f256fa00ead0f1ec6"} Nov 29 04:57:29 crc kubenswrapper[4799]: I1129 04:57:29.828436 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 29 04:57:29 crc kubenswrapper[4799]: I1129 04:57:29.874588 4799 scope.go:117] "RemoveContainer" containerID="a59906d06ff21939ff4517a38752eb41f1cd150b909901896ecad59ea285e18c" Nov 29 04:57:29 crc kubenswrapper[4799]: I1129 04:57:29.879043 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d97fcdd8f-9dgkt"] Nov 29 04:57:29 crc kubenswrapper[4799]: I1129 04:57:29.891625 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d97fcdd8f-9dgkt"] Nov 29 04:57:29 crc kubenswrapper[4799]: I1129 04:57:29.955383 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 29 04:57:29 crc kubenswrapper[4799]: I1129 04:57:29.956979 4799 scope.go:117] "RemoveContainer" containerID="ac913cd6f6b3c6cd2d97df30c634ac594219e8c013eb65754ed899e61dbaf202" Nov 29 04:57:30 crc kubenswrapper[4799]: I1129 04:57:30.003168 4799 scope.go:117] "RemoveContainer" containerID="c1d81c8095f715598cc9fb838de2ce04ef7ca5c82f6b28b2c88ab54e618a9fea" Nov 29 04:57:30 crc kubenswrapper[4799]: I1129 04:57:30.014127 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 29 04:57:30 crc kubenswrapper[4799]: I1129 04:57:30.026129 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 29 04:57:30 crc kubenswrapper[4799]: E1129 04:57:30.026607 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae86ab9e-2d28-4074-8364-9d1a01d77b62" containerName="nova-api-api" Nov 29 04:57:30 crc kubenswrapper[4799]: I1129 04:57:30.026629 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae86ab9e-2d28-4074-8364-9d1a01d77b62" containerName="nova-api-api" Nov 29 04:57:30 crc kubenswrapper[4799]: E1129 04:57:30.026658 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c419e681-9099-4422-b04e-8faccbedc127" containerName="dnsmasq-dns" Nov 29 04:57:30 crc kubenswrapper[4799]: I1129 04:57:30.026667 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="c419e681-9099-4422-b04e-8faccbedc127" containerName="dnsmasq-dns" Nov 29 04:57:30 crc kubenswrapper[4799]: E1129 04:57:30.026678 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c419e681-9099-4422-b04e-8faccbedc127" containerName="init" Nov 29 04:57:30 crc kubenswrapper[4799]: I1129 04:57:30.026684 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="c419e681-9099-4422-b04e-8faccbedc127" containerName="init" Nov 29 04:57:30 crc kubenswrapper[4799]: E1129 04:57:30.026698 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae86ab9e-2d28-4074-8364-9d1a01d77b62" containerName="nova-api-log" Nov 29 04:57:30 crc kubenswrapper[4799]: I1129 04:57:30.026704 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae86ab9e-2d28-4074-8364-9d1a01d77b62" containerName="nova-api-log" Nov 29 04:57:30 crc kubenswrapper[4799]: I1129 04:57:30.026930 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="c419e681-9099-4422-b04e-8faccbedc127" containerName="dnsmasq-dns" Nov 29 04:57:30 crc kubenswrapper[4799]: I1129 04:57:30.026947 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae86ab9e-2d28-4074-8364-9d1a01d77b62" containerName="nova-api-log" Nov 29 04:57:30 crc kubenswrapper[4799]: I1129 04:57:30.026956 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae86ab9e-2d28-4074-8364-9d1a01d77b62" containerName="nova-api-api" Nov 29 04:57:30 crc kubenswrapper[4799]: I1129 04:57:30.028297 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 29 04:57:30 crc kubenswrapper[4799]: I1129 04:57:30.031221 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 29 04:57:30 crc kubenswrapper[4799]: I1129 04:57:30.036293 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 29 04:57:30 crc kubenswrapper[4799]: I1129 04:57:30.076390 4799 scope.go:117] "RemoveContainer" containerID="ac913cd6f6b3c6cd2d97df30c634ac594219e8c013eb65754ed899e61dbaf202" Nov 29 04:57:30 crc kubenswrapper[4799]: E1129 04:57:30.077274 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac913cd6f6b3c6cd2d97df30c634ac594219e8c013eb65754ed899e61dbaf202\": container with ID starting with ac913cd6f6b3c6cd2d97df30c634ac594219e8c013eb65754ed899e61dbaf202 not found: ID does not exist" containerID="ac913cd6f6b3c6cd2d97df30c634ac594219e8c013eb65754ed899e61dbaf202" Nov 29 04:57:30 crc kubenswrapper[4799]: I1129 04:57:30.077320 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac913cd6f6b3c6cd2d97df30c634ac594219e8c013eb65754ed899e61dbaf202"} err="failed to get container status \"ac913cd6f6b3c6cd2d97df30c634ac594219e8c013eb65754ed899e61dbaf202\": rpc error: code = NotFound desc = could not find container \"ac913cd6f6b3c6cd2d97df30c634ac594219e8c013eb65754ed899e61dbaf202\": container with ID starting with ac913cd6f6b3c6cd2d97df30c634ac594219e8c013eb65754ed899e61dbaf202 not found: ID does not exist" Nov 29 04:57:30 crc kubenswrapper[4799]: I1129 04:57:30.077362 4799 scope.go:117] "RemoveContainer" containerID="c1d81c8095f715598cc9fb838de2ce04ef7ca5c82f6b28b2c88ab54e618a9fea" Nov 29 04:57:30 crc kubenswrapper[4799]: E1129 04:57:30.078545 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1d81c8095f715598cc9fb838de2ce04ef7ca5c82f6b28b2c88ab54e618a9fea\": container with ID starting with c1d81c8095f715598cc9fb838de2ce04ef7ca5c82f6b28b2c88ab54e618a9fea not found: ID does not exist" containerID="c1d81c8095f715598cc9fb838de2ce04ef7ca5c82f6b28b2c88ab54e618a9fea" Nov 29 04:57:30 crc kubenswrapper[4799]: I1129 04:57:30.078609 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1d81c8095f715598cc9fb838de2ce04ef7ca5c82f6b28b2c88ab54e618a9fea"} err="failed to get container status \"c1d81c8095f715598cc9fb838de2ce04ef7ca5c82f6b28b2c88ab54e618a9fea\": rpc error: code = NotFound desc = could not find container \"c1d81c8095f715598cc9fb838de2ce04ef7ca5c82f6b28b2c88ab54e618a9fea\": container with ID starting with c1d81c8095f715598cc9fb838de2ce04ef7ca5c82f6b28b2c88ab54e618a9fea not found: ID does not exist" Nov 29 04:57:30 crc kubenswrapper[4799]: I1129 04:57:30.207027 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4msgw\" (UniqueName: \"kubernetes.io/projected/dea67977-de74-47f9-8db0-6d032d8579c2-kube-api-access-4msgw\") pod \"nova-api-0\" (UID: \"dea67977-de74-47f9-8db0-6d032d8579c2\") " pod="openstack/nova-api-0" Nov 29 04:57:30 crc kubenswrapper[4799]: I1129 04:57:30.207714 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dea67977-de74-47f9-8db0-6d032d8579c2-logs\") pod \"nova-api-0\" (UID: \"dea67977-de74-47f9-8db0-6d032d8579c2\") " pod="openstack/nova-api-0" Nov 29 04:57:30 crc kubenswrapper[4799]: I1129 04:57:30.207841 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dea67977-de74-47f9-8db0-6d032d8579c2-config-data\") pod \"nova-api-0\" (UID: \"dea67977-de74-47f9-8db0-6d032d8579c2\") " pod="openstack/nova-api-0" Nov 29 04:57:30 crc kubenswrapper[4799]: I1129 04:57:30.207876 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dea67977-de74-47f9-8db0-6d032d8579c2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"dea67977-de74-47f9-8db0-6d032d8579c2\") " pod="openstack/nova-api-0" Nov 29 04:57:30 crc kubenswrapper[4799]: I1129 04:57:30.313519 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4msgw\" (UniqueName: \"kubernetes.io/projected/dea67977-de74-47f9-8db0-6d032d8579c2-kube-api-access-4msgw\") pod \"nova-api-0\" (UID: \"dea67977-de74-47f9-8db0-6d032d8579c2\") " pod="openstack/nova-api-0" Nov 29 04:57:30 crc kubenswrapper[4799]: I1129 04:57:30.313581 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dea67977-de74-47f9-8db0-6d032d8579c2-logs\") pod \"nova-api-0\" (UID: \"dea67977-de74-47f9-8db0-6d032d8579c2\") " pod="openstack/nova-api-0" Nov 29 04:57:30 crc kubenswrapper[4799]: I1129 04:57:30.313631 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dea67977-de74-47f9-8db0-6d032d8579c2-config-data\") pod \"nova-api-0\" (UID: \"dea67977-de74-47f9-8db0-6d032d8579c2\") " pod="openstack/nova-api-0" Nov 29 04:57:30 crc kubenswrapper[4799]: I1129 04:57:30.313663 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dea67977-de74-47f9-8db0-6d032d8579c2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"dea67977-de74-47f9-8db0-6d032d8579c2\") " pod="openstack/nova-api-0" Nov 29 04:57:30 crc kubenswrapper[4799]: I1129 04:57:30.314165 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dea67977-de74-47f9-8db0-6d032d8579c2-logs\") pod \"nova-api-0\" (UID: \"dea67977-de74-47f9-8db0-6d032d8579c2\") " pod="openstack/nova-api-0" Nov 29 04:57:30 crc kubenswrapper[4799]: I1129 04:57:30.318648 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dea67977-de74-47f9-8db0-6d032d8579c2-config-data\") pod \"nova-api-0\" (UID: \"dea67977-de74-47f9-8db0-6d032d8579c2\") " pod="openstack/nova-api-0" Nov 29 04:57:30 crc kubenswrapper[4799]: I1129 04:57:30.325066 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dea67977-de74-47f9-8db0-6d032d8579c2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"dea67977-de74-47f9-8db0-6d032d8579c2\") " pod="openstack/nova-api-0" Nov 29 04:57:30 crc kubenswrapper[4799]: I1129 04:57:30.339768 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4msgw\" (UniqueName: \"kubernetes.io/projected/dea67977-de74-47f9-8db0-6d032d8579c2-kube-api-access-4msgw\") pod \"nova-api-0\" (UID: \"dea67977-de74-47f9-8db0-6d032d8579c2\") " pod="openstack/nova-api-0" Nov 29 04:57:30 crc kubenswrapper[4799]: I1129 04:57:30.377435 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 29 04:57:30 crc kubenswrapper[4799]: I1129 04:57:30.675263 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae86ab9e-2d28-4074-8364-9d1a01d77b62" path="/var/lib/kubelet/pods/ae86ab9e-2d28-4074-8364-9d1a01d77b62/volumes" Nov 29 04:57:30 crc kubenswrapper[4799]: I1129 04:57:30.676619 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c419e681-9099-4422-b04e-8faccbedc127" path="/var/lib/kubelet/pods/c419e681-9099-4422-b04e-8faccbedc127/volumes" Nov 29 04:57:30 crc kubenswrapper[4799]: I1129 04:57:30.848134 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33323158-ffcc-4f35-9797-e1fb18c3f0f1","Type":"ContainerStarted","Data":"cd27c07c38df1fb6251e0b34585bb0d4585d48fb8bbeb278204694659344be9b"} Nov 29 04:57:30 crc kubenswrapper[4799]: I1129 04:57:30.890742 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 29 04:57:31 crc kubenswrapper[4799]: I1129 04:57:31.174940 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Nov 29 04:57:31 crc kubenswrapper[4799]: I1129 04:57:31.390696 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 29 04:57:31 crc kubenswrapper[4799]: I1129 04:57:31.545945 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06f32a7b-0ded-4927-9b8b-55fc1ecbb296-config-data\") pod \"06f32a7b-0ded-4927-9b8b-55fc1ecbb296\" (UID: \"06f32a7b-0ded-4927-9b8b-55fc1ecbb296\") " Nov 29 04:57:31 crc kubenswrapper[4799]: I1129 04:57:31.546464 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rfvz\" (UniqueName: \"kubernetes.io/projected/06f32a7b-0ded-4927-9b8b-55fc1ecbb296-kube-api-access-8rfvz\") pod \"06f32a7b-0ded-4927-9b8b-55fc1ecbb296\" (UID: \"06f32a7b-0ded-4927-9b8b-55fc1ecbb296\") " Nov 29 04:57:31 crc kubenswrapper[4799]: I1129 04:57:31.546714 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06f32a7b-0ded-4927-9b8b-55fc1ecbb296-combined-ca-bundle\") pod \"06f32a7b-0ded-4927-9b8b-55fc1ecbb296\" (UID: \"06f32a7b-0ded-4927-9b8b-55fc1ecbb296\") " Nov 29 04:57:31 crc kubenswrapper[4799]: I1129 04:57:31.557916 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06f32a7b-0ded-4927-9b8b-55fc1ecbb296-kube-api-access-8rfvz" (OuterVolumeSpecName: "kube-api-access-8rfvz") pod "06f32a7b-0ded-4927-9b8b-55fc1ecbb296" (UID: "06f32a7b-0ded-4927-9b8b-55fc1ecbb296"). InnerVolumeSpecName "kube-api-access-8rfvz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:57:31 crc kubenswrapper[4799]: I1129 04:57:31.581193 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06f32a7b-0ded-4927-9b8b-55fc1ecbb296-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "06f32a7b-0ded-4927-9b8b-55fc1ecbb296" (UID: "06f32a7b-0ded-4927-9b8b-55fc1ecbb296"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:57:31 crc kubenswrapper[4799]: I1129 04:57:31.585583 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06f32a7b-0ded-4927-9b8b-55fc1ecbb296-config-data" (OuterVolumeSpecName: "config-data") pod "06f32a7b-0ded-4927-9b8b-55fc1ecbb296" (UID: "06f32a7b-0ded-4927-9b8b-55fc1ecbb296"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:57:31 crc kubenswrapper[4799]: I1129 04:57:31.649177 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06f32a7b-0ded-4927-9b8b-55fc1ecbb296-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:31 crc kubenswrapper[4799]: I1129 04:57:31.649228 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06f32a7b-0ded-4927-9b8b-55fc1ecbb296-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:31 crc kubenswrapper[4799]: I1129 04:57:31.649240 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rfvz\" (UniqueName: \"kubernetes.io/projected/06f32a7b-0ded-4927-9b8b-55fc1ecbb296-kube-api-access-8rfvz\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:31 crc kubenswrapper[4799]: I1129 04:57:31.865245 4799 generic.go:334] "Generic (PLEG): container finished" podID="06f32a7b-0ded-4927-9b8b-55fc1ecbb296" containerID="eb674ecc4ca4399e936f237d941a66c8354d24c9fb184e868257e4e51dca7f89" exitCode=0 Nov 29 04:57:31 crc kubenswrapper[4799]: I1129 04:57:31.865326 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"06f32a7b-0ded-4927-9b8b-55fc1ecbb296","Type":"ContainerDied","Data":"eb674ecc4ca4399e936f237d941a66c8354d24c9fb184e868257e4e51dca7f89"} Nov 29 04:57:31 crc kubenswrapper[4799]: I1129 04:57:31.865361 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"06f32a7b-0ded-4927-9b8b-55fc1ecbb296","Type":"ContainerDied","Data":"31b73efb28ab805bbabc3c8844bf23a0e17e766a2449efd7457674748c96c357"} Nov 29 04:57:31 crc kubenswrapper[4799]: I1129 04:57:31.865383 4799 scope.go:117] "RemoveContainer" containerID="eb674ecc4ca4399e936f237d941a66c8354d24c9fb184e868257e4e51dca7f89" Nov 29 04:57:31 crc kubenswrapper[4799]: I1129 04:57:31.865405 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 29 04:57:31 crc kubenswrapper[4799]: I1129 04:57:31.870081 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33323158-ffcc-4f35-9797-e1fb18c3f0f1","Type":"ContainerStarted","Data":"a2060acbc541160c4825d6d930f8025b0010ff2db6b646adb9c174571f367f27"} Nov 29 04:57:31 crc kubenswrapper[4799]: I1129 04:57:31.873720 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dea67977-de74-47f9-8db0-6d032d8579c2","Type":"ContainerStarted","Data":"10c94e25337ea324516bbdcfe5f7c7d29e1e16dc5a878791e6ce18e3264e3dda"} Nov 29 04:57:31 crc kubenswrapper[4799]: I1129 04:57:31.873884 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dea67977-de74-47f9-8db0-6d032d8579c2","Type":"ContainerStarted","Data":"3f364fedaf9c50e25e9c3ced9d715beb8e2933f1fff4b185d7648a9b2b4d8a44"} Nov 29 04:57:31 crc kubenswrapper[4799]: I1129 04:57:31.873959 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dea67977-de74-47f9-8db0-6d032d8579c2","Type":"ContainerStarted","Data":"5edc1de60295dab1abfc53e0c8bd1308163896f0d5bb8931c90bda47aae8c6db"} Nov 29 04:57:31 crc kubenswrapper[4799]: I1129 04:57:31.891600 4799 scope.go:117] "RemoveContainer" containerID="eb674ecc4ca4399e936f237d941a66c8354d24c9fb184e868257e4e51dca7f89" Nov 29 04:57:31 crc kubenswrapper[4799]: E1129 04:57:31.892614 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb674ecc4ca4399e936f237d941a66c8354d24c9fb184e868257e4e51dca7f89\": container with ID starting with eb674ecc4ca4399e936f237d941a66c8354d24c9fb184e868257e4e51dca7f89 not found: ID does not exist" containerID="eb674ecc4ca4399e936f237d941a66c8354d24c9fb184e868257e4e51dca7f89" Nov 29 04:57:31 crc kubenswrapper[4799]: I1129 04:57:31.892751 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb674ecc4ca4399e936f237d941a66c8354d24c9fb184e868257e4e51dca7f89"} err="failed to get container status \"eb674ecc4ca4399e936f237d941a66c8354d24c9fb184e868257e4e51dca7f89\": rpc error: code = NotFound desc = could not find container \"eb674ecc4ca4399e936f237d941a66c8354d24c9fb184e868257e4e51dca7f89\": container with ID starting with eb674ecc4ca4399e936f237d941a66c8354d24c9fb184e868257e4e51dca7f89 not found: ID does not exist" Nov 29 04:57:31 crc kubenswrapper[4799]: I1129 04:57:31.903974 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.903953001 podStartE2EDuration="2.903953001s" podCreationTimestamp="2025-11-29 04:57:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:57:31.894957887 +0000 UTC m=+1127.537888287" watchObservedRunningTime="2025-11-29 04:57:31.903953001 +0000 UTC m=+1127.546883401" Nov 29 04:57:31 crc kubenswrapper[4799]: I1129 04:57:31.932463 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 29 04:57:31 crc kubenswrapper[4799]: I1129 04:57:31.937700 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 29 04:57:31 crc kubenswrapper[4799]: I1129 04:57:31.950146 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 29 04:57:31 crc kubenswrapper[4799]: E1129 04:57:31.950747 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06f32a7b-0ded-4927-9b8b-55fc1ecbb296" containerName="nova-scheduler-scheduler" Nov 29 04:57:31 crc kubenswrapper[4799]: I1129 04:57:31.950779 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="06f32a7b-0ded-4927-9b8b-55fc1ecbb296" containerName="nova-scheduler-scheduler" Nov 29 04:57:31 crc kubenswrapper[4799]: I1129 04:57:31.951032 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="06f32a7b-0ded-4927-9b8b-55fc1ecbb296" containerName="nova-scheduler-scheduler" Nov 29 04:57:31 crc kubenswrapper[4799]: I1129 04:57:31.952066 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 29 04:57:31 crc kubenswrapper[4799]: I1129 04:57:31.954860 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 29 04:57:31 crc kubenswrapper[4799]: I1129 04:57:31.968510 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 29 04:57:32 crc kubenswrapper[4799]: I1129 04:57:32.063102 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c25c970-ef30-438b-b5d3-a98976fa7caa-config-data\") pod \"nova-scheduler-0\" (UID: \"3c25c970-ef30-438b-b5d3-a98976fa7caa\") " pod="openstack/nova-scheduler-0" Nov 29 04:57:32 crc kubenswrapper[4799]: I1129 04:57:32.063277 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c25c970-ef30-438b-b5d3-a98976fa7caa-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"3c25c970-ef30-438b-b5d3-a98976fa7caa\") " pod="openstack/nova-scheduler-0" Nov 29 04:57:32 crc kubenswrapper[4799]: I1129 04:57:32.063387 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhknj\" (UniqueName: \"kubernetes.io/projected/3c25c970-ef30-438b-b5d3-a98976fa7caa-kube-api-access-zhknj\") pod \"nova-scheduler-0\" (UID: \"3c25c970-ef30-438b-b5d3-a98976fa7caa\") " pod="openstack/nova-scheduler-0" Nov 29 04:57:32 crc kubenswrapper[4799]: I1129 04:57:32.165479 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhknj\" (UniqueName: \"kubernetes.io/projected/3c25c970-ef30-438b-b5d3-a98976fa7caa-kube-api-access-zhknj\") pod \"nova-scheduler-0\" (UID: \"3c25c970-ef30-438b-b5d3-a98976fa7caa\") " pod="openstack/nova-scheduler-0" Nov 29 04:57:32 crc kubenswrapper[4799]: I1129 04:57:32.166081 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c25c970-ef30-438b-b5d3-a98976fa7caa-config-data\") pod \"nova-scheduler-0\" (UID: \"3c25c970-ef30-438b-b5d3-a98976fa7caa\") " pod="openstack/nova-scheduler-0" Nov 29 04:57:32 crc kubenswrapper[4799]: I1129 04:57:32.166448 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c25c970-ef30-438b-b5d3-a98976fa7caa-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"3c25c970-ef30-438b-b5d3-a98976fa7caa\") " pod="openstack/nova-scheduler-0" Nov 29 04:57:32 crc kubenswrapper[4799]: I1129 04:57:32.179300 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c25c970-ef30-438b-b5d3-a98976fa7caa-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"3c25c970-ef30-438b-b5d3-a98976fa7caa\") " pod="openstack/nova-scheduler-0" Nov 29 04:57:32 crc kubenswrapper[4799]: I1129 04:57:32.180682 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c25c970-ef30-438b-b5d3-a98976fa7caa-config-data\") pod \"nova-scheduler-0\" (UID: \"3c25c970-ef30-438b-b5d3-a98976fa7caa\") " pod="openstack/nova-scheduler-0" Nov 29 04:57:32 crc kubenswrapper[4799]: I1129 04:57:32.187140 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhknj\" (UniqueName: \"kubernetes.io/projected/3c25c970-ef30-438b-b5d3-a98976fa7caa-kube-api-access-zhknj\") pod \"nova-scheduler-0\" (UID: \"3c25c970-ef30-438b-b5d3-a98976fa7caa\") " pod="openstack/nova-scheduler-0" Nov 29 04:57:32 crc kubenswrapper[4799]: I1129 04:57:32.314809 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 29 04:57:32 crc kubenswrapper[4799]: I1129 04:57:32.698579 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06f32a7b-0ded-4927-9b8b-55fc1ecbb296" path="/var/lib/kubelet/pods/06f32a7b-0ded-4927-9b8b-55fc1ecbb296/volumes" Nov 29 04:57:32 crc kubenswrapper[4799]: I1129 04:57:32.829357 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 29 04:57:32 crc kubenswrapper[4799]: W1129 04:57:32.832810 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c25c970_ef30_438b_b5d3_a98976fa7caa.slice/crio-551552a170f802f9a09bcba020727ba2583db85271234e8c871ca4157602c855 WatchSource:0}: Error finding container 551552a170f802f9a09bcba020727ba2583db85271234e8c871ca4157602c855: Status 404 returned error can't find the container with id 551552a170f802f9a09bcba020727ba2583db85271234e8c871ca4157602c855 Nov 29 04:57:32 crc kubenswrapper[4799]: I1129 04:57:32.883335 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3c25c970-ef30-438b-b5d3-a98976fa7caa","Type":"ContainerStarted","Data":"551552a170f802f9a09bcba020727ba2583db85271234e8c871ca4157602c855"} Nov 29 04:57:32 crc kubenswrapper[4799]: I1129 04:57:32.885400 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33323158-ffcc-4f35-9797-e1fb18c3f0f1","Type":"ContainerStarted","Data":"ddabe9f210f448ee6eb45e8f616342f38b8ed24a13d6a5a5e213be02cd5fec35"} Nov 29 04:57:32 crc kubenswrapper[4799]: I1129 04:57:32.885572 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 29 04:57:32 crc kubenswrapper[4799]: I1129 04:57:32.913622 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.355531355 podStartE2EDuration="5.913577768s" podCreationTimestamp="2025-11-29 04:57:27 +0000 UTC" firstStartedPulling="2025-11-29 04:57:28.757311769 +0000 UTC m=+1124.400242169" lastFinishedPulling="2025-11-29 04:57:32.315358182 +0000 UTC m=+1127.958288582" observedRunningTime="2025-11-29 04:57:32.903654772 +0000 UTC m=+1128.546585172" watchObservedRunningTime="2025-11-29 04:57:32.913577768 +0000 UTC m=+1128.556508168" Nov 29 04:57:33 crc kubenswrapper[4799]: I1129 04:57:33.100838 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 29 04:57:33 crc kubenswrapper[4799]: I1129 04:57:33.899865 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3c25c970-ef30-438b-b5d3-a98976fa7caa","Type":"ContainerStarted","Data":"1879026d9dc12823fb9939d498c22c08936c600c692b8f34e8b5385e53f855f4"} Nov 29 04:57:33 crc kubenswrapper[4799]: I1129 04:57:33.924887 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.924862014 podStartE2EDuration="2.924862014s" podCreationTimestamp="2025-11-29 04:57:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:57:33.919877296 +0000 UTC m=+1129.562807696" watchObservedRunningTime="2025-11-29 04:57:33.924862014 +0000 UTC m=+1129.567792434" Nov 29 04:57:37 crc kubenswrapper[4799]: I1129 04:57:37.315983 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 29 04:57:37 crc kubenswrapper[4799]: I1129 04:57:37.497944 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 04:57:37 crc kubenswrapper[4799]: I1129 04:57:37.498046 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 04:57:40 crc kubenswrapper[4799]: I1129 04:57:40.377943 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 29 04:57:40 crc kubenswrapper[4799]: I1129 04:57:40.378404 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 29 04:57:41 crc kubenswrapper[4799]: I1129 04:57:41.460077 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="dea67977-de74-47f9-8db0-6d032d8579c2" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.175:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 29 04:57:41 crc kubenswrapper[4799]: I1129 04:57:41.460083 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="dea67977-de74-47f9-8db0-6d032d8579c2" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.175:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 29 04:57:42 crc kubenswrapper[4799]: I1129 04:57:42.315985 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 29 04:57:42 crc kubenswrapper[4799]: I1129 04:57:42.346289 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 29 04:57:43 crc kubenswrapper[4799]: I1129 04:57:43.019263 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 29 04:57:50 crc kubenswrapper[4799]: I1129 04:57:50.062764 4799 generic.go:334] "Generic (PLEG): container finished" podID="3603071d-80f3-4d51-983f-282a0fb33ac4" containerID="95d7d63500012b1fa8cd80b35d7674be5e6fe32e79244a1d378e9daf4ce43303" exitCode=137 Nov 29 04:57:50 crc kubenswrapper[4799]: I1129 04:57:50.062860 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3603071d-80f3-4d51-983f-282a0fb33ac4","Type":"ContainerDied","Data":"95d7d63500012b1fa8cd80b35d7674be5e6fe32e79244a1d378e9daf4ce43303"} Nov 29 04:57:50 crc kubenswrapper[4799]: I1129 04:57:50.066848 4799 generic.go:334] "Generic (PLEG): container finished" podID="68fa593e-4704-40d7-a86d-ae984e5129cc" containerID="cfeb3da10847f7ebeda995678cedbbba4c0edf34e678a1a155acb97431284903" exitCode=137 Nov 29 04:57:50 crc kubenswrapper[4799]: I1129 04:57:50.066902 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"68fa593e-4704-40d7-a86d-ae984e5129cc","Type":"ContainerDied","Data":"cfeb3da10847f7ebeda995678cedbbba4c0edf34e678a1a155acb97431284903"} Nov 29 04:57:50 crc kubenswrapper[4799]: I1129 04:57:50.383711 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 29 04:57:50 crc kubenswrapper[4799]: I1129 04:57:50.384253 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 29 04:57:50 crc kubenswrapper[4799]: I1129 04:57:50.389400 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 29 04:57:50 crc kubenswrapper[4799]: I1129 04:57:50.389800 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 29 04:57:50 crc kubenswrapper[4799]: I1129 04:57:50.563078 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 29 04:57:50 crc kubenswrapper[4799]: I1129 04:57:50.569614 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 29 04:57:50 crc kubenswrapper[4799]: I1129 04:57:50.651016 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68fa593e-4704-40d7-a86d-ae984e5129cc-logs\") pod \"68fa593e-4704-40d7-a86d-ae984e5129cc\" (UID: \"68fa593e-4704-40d7-a86d-ae984e5129cc\") " Nov 29 04:57:50 crc kubenswrapper[4799]: I1129 04:57:50.651115 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3603071d-80f3-4d51-983f-282a0fb33ac4-combined-ca-bundle\") pod \"3603071d-80f3-4d51-983f-282a0fb33ac4\" (UID: \"3603071d-80f3-4d51-983f-282a0fb33ac4\") " Nov 29 04:57:50 crc kubenswrapper[4799]: I1129 04:57:50.651212 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68fa593e-4704-40d7-a86d-ae984e5129cc-config-data\") pod \"68fa593e-4704-40d7-a86d-ae984e5129cc\" (UID: \"68fa593e-4704-40d7-a86d-ae984e5129cc\") " Nov 29 04:57:50 crc kubenswrapper[4799]: I1129 04:57:50.651320 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3603071d-80f3-4d51-983f-282a0fb33ac4-config-data\") pod \"3603071d-80f3-4d51-983f-282a0fb33ac4\" (UID: \"3603071d-80f3-4d51-983f-282a0fb33ac4\") " Nov 29 04:57:50 crc kubenswrapper[4799]: I1129 04:57:50.651349 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbpw9\" (UniqueName: \"kubernetes.io/projected/3603071d-80f3-4d51-983f-282a0fb33ac4-kube-api-access-bbpw9\") pod \"3603071d-80f3-4d51-983f-282a0fb33ac4\" (UID: \"3603071d-80f3-4d51-983f-282a0fb33ac4\") " Nov 29 04:57:50 crc kubenswrapper[4799]: I1129 04:57:50.651383 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68fa593e-4704-40d7-a86d-ae984e5129cc-combined-ca-bundle\") pod \"68fa593e-4704-40d7-a86d-ae984e5129cc\" (UID: \"68fa593e-4704-40d7-a86d-ae984e5129cc\") " Nov 29 04:57:50 crc kubenswrapper[4799]: I1129 04:57:50.651422 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-88rdm\" (UniqueName: \"kubernetes.io/projected/68fa593e-4704-40d7-a86d-ae984e5129cc-kube-api-access-88rdm\") pod \"68fa593e-4704-40d7-a86d-ae984e5129cc\" (UID: \"68fa593e-4704-40d7-a86d-ae984e5129cc\") " Nov 29 04:57:50 crc kubenswrapper[4799]: I1129 04:57:50.652486 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68fa593e-4704-40d7-a86d-ae984e5129cc-logs" (OuterVolumeSpecName: "logs") pod "68fa593e-4704-40d7-a86d-ae984e5129cc" (UID: "68fa593e-4704-40d7-a86d-ae984e5129cc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:57:50 crc kubenswrapper[4799]: I1129 04:57:50.658971 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68fa593e-4704-40d7-a86d-ae984e5129cc-kube-api-access-88rdm" (OuterVolumeSpecName: "kube-api-access-88rdm") pod "68fa593e-4704-40d7-a86d-ae984e5129cc" (UID: "68fa593e-4704-40d7-a86d-ae984e5129cc"). InnerVolumeSpecName "kube-api-access-88rdm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:57:50 crc kubenswrapper[4799]: I1129 04:57:50.661986 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3603071d-80f3-4d51-983f-282a0fb33ac4-kube-api-access-bbpw9" (OuterVolumeSpecName: "kube-api-access-bbpw9") pod "3603071d-80f3-4d51-983f-282a0fb33ac4" (UID: "3603071d-80f3-4d51-983f-282a0fb33ac4"). InnerVolumeSpecName "kube-api-access-bbpw9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:57:50 crc kubenswrapper[4799]: I1129 04:57:50.693259 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3603071d-80f3-4d51-983f-282a0fb33ac4-config-data" (OuterVolumeSpecName: "config-data") pod "3603071d-80f3-4d51-983f-282a0fb33ac4" (UID: "3603071d-80f3-4d51-983f-282a0fb33ac4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:57:50 crc kubenswrapper[4799]: I1129 04:57:50.694469 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68fa593e-4704-40d7-a86d-ae984e5129cc-config-data" (OuterVolumeSpecName: "config-data") pod "68fa593e-4704-40d7-a86d-ae984e5129cc" (UID: "68fa593e-4704-40d7-a86d-ae984e5129cc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:57:50 crc kubenswrapper[4799]: I1129 04:57:50.694545 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3603071d-80f3-4d51-983f-282a0fb33ac4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3603071d-80f3-4d51-983f-282a0fb33ac4" (UID: "3603071d-80f3-4d51-983f-282a0fb33ac4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:57:50 crc kubenswrapper[4799]: I1129 04:57:50.694961 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68fa593e-4704-40d7-a86d-ae984e5129cc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "68fa593e-4704-40d7-a86d-ae984e5129cc" (UID: "68fa593e-4704-40d7-a86d-ae984e5129cc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:57:50 crc kubenswrapper[4799]: I1129 04:57:50.753328 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3603071d-80f3-4d51-983f-282a0fb33ac4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:50 crc kubenswrapper[4799]: I1129 04:57:50.753361 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68fa593e-4704-40d7-a86d-ae984e5129cc-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:50 crc kubenswrapper[4799]: I1129 04:57:50.753370 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3603071d-80f3-4d51-983f-282a0fb33ac4-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:50 crc kubenswrapper[4799]: I1129 04:57:50.753378 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbpw9\" (UniqueName: \"kubernetes.io/projected/3603071d-80f3-4d51-983f-282a0fb33ac4-kube-api-access-bbpw9\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:50 crc kubenswrapper[4799]: I1129 04:57:50.753389 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68fa593e-4704-40d7-a86d-ae984e5129cc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:50 crc kubenswrapper[4799]: I1129 04:57:50.753397 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-88rdm\" (UniqueName: \"kubernetes.io/projected/68fa593e-4704-40d7-a86d-ae984e5129cc-kube-api-access-88rdm\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:50 crc kubenswrapper[4799]: I1129 04:57:50.753406 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68fa593e-4704-40d7-a86d-ae984e5129cc-logs\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.078272 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.078288 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3603071d-80f3-4d51-983f-282a0fb33ac4","Type":"ContainerDied","Data":"2958e40b129c8a386f93b97faedd3136b51b07eae64ea53be56627f207f43e93"} Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.078539 4799 scope.go:117] "RemoveContainer" containerID="95d7d63500012b1fa8cd80b35d7674be5e6fe32e79244a1d378e9daf4ce43303" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.081000 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"68fa593e-4704-40d7-a86d-ae984e5129cc","Type":"ContainerDied","Data":"18ff036b3ca2b7c2caa69921456d2aafa72c8241293c8f4e760ad8718d058a19"} Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.081011 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.081290 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.086680 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.106748 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.108543 4799 scope.go:117] "RemoveContainer" containerID="cfeb3da10847f7ebeda995678cedbbba4c0edf34e678a1a155acb97431284903" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.118428 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.134768 4799 scope.go:117] "RemoveContainer" containerID="003f5f407757a793f711e60be6c4870f1a4a1e1cbe2647fdf10cc619f51a2e01" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.152897 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.173931 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.185343 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 29 04:57:51 crc kubenswrapper[4799]: E1129 04:57:51.185773 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3603071d-80f3-4d51-983f-282a0fb33ac4" containerName="nova-cell1-novncproxy-novncproxy" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.185866 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="3603071d-80f3-4d51-983f-282a0fb33ac4" containerName="nova-cell1-novncproxy-novncproxy" Nov 29 04:57:51 crc kubenswrapper[4799]: E1129 04:57:51.185887 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68fa593e-4704-40d7-a86d-ae984e5129cc" containerName="nova-metadata-log" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.185895 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="68fa593e-4704-40d7-a86d-ae984e5129cc" containerName="nova-metadata-log" Nov 29 04:57:51 crc kubenswrapper[4799]: E1129 04:57:51.185902 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68fa593e-4704-40d7-a86d-ae984e5129cc" containerName="nova-metadata-metadata" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.185909 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="68fa593e-4704-40d7-a86d-ae984e5129cc" containerName="nova-metadata-metadata" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.186105 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="68fa593e-4704-40d7-a86d-ae984e5129cc" containerName="nova-metadata-log" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.186117 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="3603071d-80f3-4d51-983f-282a0fb33ac4" containerName="nova-cell1-novncproxy-novncproxy" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.186132 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="68fa593e-4704-40d7-a86d-ae984e5129cc" containerName="nova-metadata-metadata" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.187026 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.189391 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.189573 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.189697 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.207955 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.209650 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.215241 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.215813 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.252925 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.262094 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/91d09977-e990-4df4-921c-6f087e18b85f-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"91d09977-e990-4df4-921c-6f087e18b85f\") " pod="openstack/nova-cell1-novncproxy-0" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.262143 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e2d8276-3c6c-418a-8d82-7b574dc1dd63-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4e2d8276-3c6c-418a-8d82-7b574dc1dd63\") " pod="openstack/nova-metadata-0" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.262267 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e2d8276-3c6c-418a-8d82-7b574dc1dd63-config-data\") pod \"nova-metadata-0\" (UID: \"4e2d8276-3c6c-418a-8d82-7b574dc1dd63\") " pod="openstack/nova-metadata-0" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.262354 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vmq9\" (UniqueName: \"kubernetes.io/projected/91d09977-e990-4df4-921c-6f087e18b85f-kube-api-access-4vmq9\") pod \"nova-cell1-novncproxy-0\" (UID: \"91d09977-e990-4df4-921c-6f087e18b85f\") " pod="openstack/nova-cell1-novncproxy-0" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.262379 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e2d8276-3c6c-418a-8d82-7b574dc1dd63-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4e2d8276-3c6c-418a-8d82-7b574dc1dd63\") " pod="openstack/nova-metadata-0" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.262419 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/91d09977-e990-4df4-921c-6f087e18b85f-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"91d09977-e990-4df4-921c-6f087e18b85f\") " pod="openstack/nova-cell1-novncproxy-0" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.262476 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xb8rj\" (UniqueName: \"kubernetes.io/projected/4e2d8276-3c6c-418a-8d82-7b574dc1dd63-kube-api-access-xb8rj\") pod \"nova-metadata-0\" (UID: \"4e2d8276-3c6c-418a-8d82-7b574dc1dd63\") " pod="openstack/nova-metadata-0" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.262606 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91d09977-e990-4df4-921c-6f087e18b85f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"91d09977-e990-4df4-921c-6f087e18b85f\") " pod="openstack/nova-cell1-novncproxy-0" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.262657 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e2d8276-3c6c-418a-8d82-7b574dc1dd63-logs\") pod \"nova-metadata-0\" (UID: \"4e2d8276-3c6c-418a-8d82-7b574dc1dd63\") " pod="openstack/nova-metadata-0" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.262706 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91d09977-e990-4df4-921c-6f087e18b85f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"91d09977-e990-4df4-921c-6f087e18b85f\") " pod="openstack/nova-cell1-novncproxy-0" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.266232 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.328942 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b856c5697-jgvwh"] Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.330499 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b856c5697-jgvwh" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.360566 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b856c5697-jgvwh"] Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.367214 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e2d8276-3c6c-418a-8d82-7b574dc1dd63-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4e2d8276-3c6c-418a-8d82-7b574dc1dd63\") " pod="openstack/nova-metadata-0" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.367271 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hl42r\" (UniqueName: \"kubernetes.io/projected/c0b29191-5d29-44fa-9cde-c2604b5426bb-kube-api-access-hl42r\") pod \"dnsmasq-dns-5b856c5697-jgvwh\" (UID: \"c0b29191-5d29-44fa-9cde-c2604b5426bb\") " pod="openstack/dnsmasq-dns-5b856c5697-jgvwh" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.367303 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e2d8276-3c6c-418a-8d82-7b574dc1dd63-config-data\") pod \"nova-metadata-0\" (UID: \"4e2d8276-3c6c-418a-8d82-7b574dc1dd63\") " pod="openstack/nova-metadata-0" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.367362 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vmq9\" (UniqueName: \"kubernetes.io/projected/91d09977-e990-4df4-921c-6f087e18b85f-kube-api-access-4vmq9\") pod \"nova-cell1-novncproxy-0\" (UID: \"91d09977-e990-4df4-921c-6f087e18b85f\") " pod="openstack/nova-cell1-novncproxy-0" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.367382 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e2d8276-3c6c-418a-8d82-7b574dc1dd63-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4e2d8276-3c6c-418a-8d82-7b574dc1dd63\") " pod="openstack/nova-metadata-0" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.367407 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0b29191-5d29-44fa-9cde-c2604b5426bb-config\") pod \"dnsmasq-dns-5b856c5697-jgvwh\" (UID: \"c0b29191-5d29-44fa-9cde-c2604b5426bb\") " pod="openstack/dnsmasq-dns-5b856c5697-jgvwh" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.367425 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c0b29191-5d29-44fa-9cde-c2604b5426bb-ovsdbserver-nb\") pod \"dnsmasq-dns-5b856c5697-jgvwh\" (UID: \"c0b29191-5d29-44fa-9cde-c2604b5426bb\") " pod="openstack/dnsmasq-dns-5b856c5697-jgvwh" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.367454 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/91d09977-e990-4df4-921c-6f087e18b85f-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"91d09977-e990-4df4-921c-6f087e18b85f\") " pod="openstack/nova-cell1-novncproxy-0" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.367512 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xb8rj\" (UniqueName: \"kubernetes.io/projected/4e2d8276-3c6c-418a-8d82-7b574dc1dd63-kube-api-access-xb8rj\") pod \"nova-metadata-0\" (UID: \"4e2d8276-3c6c-418a-8d82-7b574dc1dd63\") " pod="openstack/nova-metadata-0" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.367625 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91d09977-e990-4df4-921c-6f087e18b85f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"91d09977-e990-4df4-921c-6f087e18b85f\") " pod="openstack/nova-cell1-novncproxy-0" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.367668 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e2d8276-3c6c-418a-8d82-7b574dc1dd63-logs\") pod \"nova-metadata-0\" (UID: \"4e2d8276-3c6c-418a-8d82-7b574dc1dd63\") " pod="openstack/nova-metadata-0" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.367692 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c0b29191-5d29-44fa-9cde-c2604b5426bb-ovsdbserver-sb\") pod \"dnsmasq-dns-5b856c5697-jgvwh\" (UID: \"c0b29191-5d29-44fa-9cde-c2604b5426bb\") " pod="openstack/dnsmasq-dns-5b856c5697-jgvwh" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.367728 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c0b29191-5d29-44fa-9cde-c2604b5426bb-dns-svc\") pod \"dnsmasq-dns-5b856c5697-jgvwh\" (UID: \"c0b29191-5d29-44fa-9cde-c2604b5426bb\") " pod="openstack/dnsmasq-dns-5b856c5697-jgvwh" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.367753 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91d09977-e990-4df4-921c-6f087e18b85f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"91d09977-e990-4df4-921c-6f087e18b85f\") " pod="openstack/nova-cell1-novncproxy-0" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.367922 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/91d09977-e990-4df4-921c-6f087e18b85f-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"91d09977-e990-4df4-921c-6f087e18b85f\") " pod="openstack/nova-cell1-novncproxy-0" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.368361 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e2d8276-3c6c-418a-8d82-7b574dc1dd63-logs\") pod \"nova-metadata-0\" (UID: \"4e2d8276-3c6c-418a-8d82-7b574dc1dd63\") " pod="openstack/nova-metadata-0" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.373459 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e2d8276-3c6c-418a-8d82-7b574dc1dd63-config-data\") pod \"nova-metadata-0\" (UID: \"4e2d8276-3c6c-418a-8d82-7b574dc1dd63\") " pod="openstack/nova-metadata-0" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.382635 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/91d09977-e990-4df4-921c-6f087e18b85f-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"91d09977-e990-4df4-921c-6f087e18b85f\") " pod="openstack/nova-cell1-novncproxy-0" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.383343 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91d09977-e990-4df4-921c-6f087e18b85f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"91d09977-e990-4df4-921c-6f087e18b85f\") " pod="openstack/nova-cell1-novncproxy-0" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.383530 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e2d8276-3c6c-418a-8d82-7b574dc1dd63-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4e2d8276-3c6c-418a-8d82-7b574dc1dd63\") " pod="openstack/nova-metadata-0" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.383723 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/91d09977-e990-4df4-921c-6f087e18b85f-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"91d09977-e990-4df4-921c-6f087e18b85f\") " pod="openstack/nova-cell1-novncproxy-0" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.384158 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e2d8276-3c6c-418a-8d82-7b574dc1dd63-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4e2d8276-3c6c-418a-8d82-7b574dc1dd63\") " pod="openstack/nova-metadata-0" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.389535 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vmq9\" (UniqueName: \"kubernetes.io/projected/91d09977-e990-4df4-921c-6f087e18b85f-kube-api-access-4vmq9\") pod \"nova-cell1-novncproxy-0\" (UID: \"91d09977-e990-4df4-921c-6f087e18b85f\") " pod="openstack/nova-cell1-novncproxy-0" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.391107 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xb8rj\" (UniqueName: \"kubernetes.io/projected/4e2d8276-3c6c-418a-8d82-7b574dc1dd63-kube-api-access-xb8rj\") pod \"nova-metadata-0\" (UID: \"4e2d8276-3c6c-418a-8d82-7b574dc1dd63\") " pod="openstack/nova-metadata-0" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.391376 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91d09977-e990-4df4-921c-6f087e18b85f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"91d09977-e990-4df4-921c-6f087e18b85f\") " pod="openstack/nova-cell1-novncproxy-0" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.469683 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c0b29191-5d29-44fa-9cde-c2604b5426bb-ovsdbserver-sb\") pod \"dnsmasq-dns-5b856c5697-jgvwh\" (UID: \"c0b29191-5d29-44fa-9cde-c2604b5426bb\") " pod="openstack/dnsmasq-dns-5b856c5697-jgvwh" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.470141 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c0b29191-5d29-44fa-9cde-c2604b5426bb-dns-svc\") pod \"dnsmasq-dns-5b856c5697-jgvwh\" (UID: \"c0b29191-5d29-44fa-9cde-c2604b5426bb\") " pod="openstack/dnsmasq-dns-5b856c5697-jgvwh" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.470216 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hl42r\" (UniqueName: \"kubernetes.io/projected/c0b29191-5d29-44fa-9cde-c2604b5426bb-kube-api-access-hl42r\") pod \"dnsmasq-dns-5b856c5697-jgvwh\" (UID: \"c0b29191-5d29-44fa-9cde-c2604b5426bb\") " pod="openstack/dnsmasq-dns-5b856c5697-jgvwh" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.470250 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0b29191-5d29-44fa-9cde-c2604b5426bb-config\") pod \"dnsmasq-dns-5b856c5697-jgvwh\" (UID: \"c0b29191-5d29-44fa-9cde-c2604b5426bb\") " pod="openstack/dnsmasq-dns-5b856c5697-jgvwh" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.470271 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c0b29191-5d29-44fa-9cde-c2604b5426bb-ovsdbserver-nb\") pod \"dnsmasq-dns-5b856c5697-jgvwh\" (UID: \"c0b29191-5d29-44fa-9cde-c2604b5426bb\") " pod="openstack/dnsmasq-dns-5b856c5697-jgvwh" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.470711 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c0b29191-5d29-44fa-9cde-c2604b5426bb-ovsdbserver-sb\") pod \"dnsmasq-dns-5b856c5697-jgvwh\" (UID: \"c0b29191-5d29-44fa-9cde-c2604b5426bb\") " pod="openstack/dnsmasq-dns-5b856c5697-jgvwh" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.471077 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c0b29191-5d29-44fa-9cde-c2604b5426bb-ovsdbserver-nb\") pod \"dnsmasq-dns-5b856c5697-jgvwh\" (UID: \"c0b29191-5d29-44fa-9cde-c2604b5426bb\") " pod="openstack/dnsmasq-dns-5b856c5697-jgvwh" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.471513 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c0b29191-5d29-44fa-9cde-c2604b5426bb-dns-svc\") pod \"dnsmasq-dns-5b856c5697-jgvwh\" (UID: \"c0b29191-5d29-44fa-9cde-c2604b5426bb\") " pod="openstack/dnsmasq-dns-5b856c5697-jgvwh" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.472420 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0b29191-5d29-44fa-9cde-c2604b5426bb-config\") pod \"dnsmasq-dns-5b856c5697-jgvwh\" (UID: \"c0b29191-5d29-44fa-9cde-c2604b5426bb\") " pod="openstack/dnsmasq-dns-5b856c5697-jgvwh" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.487939 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hl42r\" (UniqueName: \"kubernetes.io/projected/c0b29191-5d29-44fa-9cde-c2604b5426bb-kube-api-access-hl42r\") pod \"dnsmasq-dns-5b856c5697-jgvwh\" (UID: \"c0b29191-5d29-44fa-9cde-c2604b5426bb\") " pod="openstack/dnsmasq-dns-5b856c5697-jgvwh" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.516166 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.529151 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 29 04:57:51 crc kubenswrapper[4799]: I1129 04:57:51.653619 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b856c5697-jgvwh" Nov 29 04:57:52 crc kubenswrapper[4799]: I1129 04:57:52.095415 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 29 04:57:52 crc kubenswrapper[4799]: I1129 04:57:52.175897 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 29 04:57:52 crc kubenswrapper[4799]: I1129 04:57:52.262761 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b856c5697-jgvwh"] Nov 29 04:57:52 crc kubenswrapper[4799]: W1129 04:57:52.272469 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0b29191_5d29_44fa_9cde_c2604b5426bb.slice/crio-15e83276f8610bc12c801e4748f4a50e278929c7450ae3f6c9c44af7e9a9fbf9 WatchSource:0}: Error finding container 15e83276f8610bc12c801e4748f4a50e278929c7450ae3f6c9c44af7e9a9fbf9: Status 404 returned error can't find the container with id 15e83276f8610bc12c801e4748f4a50e278929c7450ae3f6c9c44af7e9a9fbf9 Nov 29 04:57:52 crc kubenswrapper[4799]: I1129 04:57:52.669689 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3603071d-80f3-4d51-983f-282a0fb33ac4" path="/var/lib/kubelet/pods/3603071d-80f3-4d51-983f-282a0fb33ac4/volumes" Nov 29 04:57:52 crc kubenswrapper[4799]: I1129 04:57:52.670509 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68fa593e-4704-40d7-a86d-ae984e5129cc" path="/var/lib/kubelet/pods/68fa593e-4704-40d7-a86d-ae984e5129cc/volumes" Nov 29 04:57:53 crc kubenswrapper[4799]: I1129 04:57:53.113323 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4e2d8276-3c6c-418a-8d82-7b574dc1dd63","Type":"ContainerStarted","Data":"38b6e6f85723c79313b1275c8ce3a65d703aa753ec7fb78e4971491a0cce0102"} Nov 29 04:57:53 crc kubenswrapper[4799]: I1129 04:57:53.113607 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4e2d8276-3c6c-418a-8d82-7b574dc1dd63","Type":"ContainerStarted","Data":"ae0ad8cc7e6278c605e29f443008d725401ca68b95af91b61cd7a19cbef8fa66"} Nov 29 04:57:53 crc kubenswrapper[4799]: I1129 04:57:53.113621 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4e2d8276-3c6c-418a-8d82-7b574dc1dd63","Type":"ContainerStarted","Data":"2a43d4c7c1590b1a1006eb0d7782448217401cb2fee1ec735600f117684db004"} Nov 29 04:57:53 crc kubenswrapper[4799]: I1129 04:57:53.115698 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"91d09977-e990-4df4-921c-6f087e18b85f","Type":"ContainerStarted","Data":"e5ce7705c711f738e381dd90d28b8328b10524d648bff8edd8ffe0ac63172516"} Nov 29 04:57:53 crc kubenswrapper[4799]: I1129 04:57:53.115753 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"91d09977-e990-4df4-921c-6f087e18b85f","Type":"ContainerStarted","Data":"282cedb6e88e544c92a36d3463de30ada31a2b366222653db3e7d796686ec855"} Nov 29 04:57:53 crc kubenswrapper[4799]: I1129 04:57:53.120209 4799 generic.go:334] "Generic (PLEG): container finished" podID="c0b29191-5d29-44fa-9cde-c2604b5426bb" containerID="230e52c4037b2f744e5c683c4c87ce6888eeef33464b66a4c1252f5b3054cf7e" exitCode=0 Nov 29 04:57:53 crc kubenswrapper[4799]: I1129 04:57:53.121030 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b856c5697-jgvwh" event={"ID":"c0b29191-5d29-44fa-9cde-c2604b5426bb","Type":"ContainerDied","Data":"230e52c4037b2f744e5c683c4c87ce6888eeef33464b66a4c1252f5b3054cf7e"} Nov 29 04:57:53 crc kubenswrapper[4799]: I1129 04:57:53.121067 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b856c5697-jgvwh" event={"ID":"c0b29191-5d29-44fa-9cde-c2604b5426bb","Type":"ContainerStarted","Data":"15e83276f8610bc12c801e4748f4a50e278929c7450ae3f6c9c44af7e9a9fbf9"} Nov 29 04:57:53 crc kubenswrapper[4799]: I1129 04:57:53.152115 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.152087733 podStartE2EDuration="2.152087733s" podCreationTimestamp="2025-11-29 04:57:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:57:53.137844855 +0000 UTC m=+1148.780775265" watchObservedRunningTime="2025-11-29 04:57:53.152087733 +0000 UTC m=+1148.795018133" Nov 29 04:57:53 crc kubenswrapper[4799]: I1129 04:57:53.203269 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.203241481 podStartE2EDuration="2.203241481s" podCreationTimestamp="2025-11-29 04:57:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:57:53.195887186 +0000 UTC m=+1148.838817616" watchObservedRunningTime="2025-11-29 04:57:53.203241481 +0000 UTC m=+1148.846171881" Nov 29 04:57:53 crc kubenswrapper[4799]: I1129 04:57:53.779536 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 29 04:57:53 crc kubenswrapper[4799]: I1129 04:57:53.780331 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="33323158-ffcc-4f35-9797-e1fb18c3f0f1" containerName="ceilometer-central-agent" containerID="cri-o://9d0c07f09a3c0e24fb8bfd28d5f8c8ee44af9d249f5f1d1fb9b8f47a7f779370" gracePeriod=30 Nov 29 04:57:53 crc kubenswrapper[4799]: I1129 04:57:53.780505 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="33323158-ffcc-4f35-9797-e1fb18c3f0f1" containerName="proxy-httpd" containerID="cri-o://ddabe9f210f448ee6eb45e8f616342f38b8ed24a13d6a5a5e213be02cd5fec35" gracePeriod=30 Nov 29 04:57:53 crc kubenswrapper[4799]: I1129 04:57:53.780570 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="33323158-ffcc-4f35-9797-e1fb18c3f0f1" containerName="sg-core" containerID="cri-o://a2060acbc541160c4825d6d930f8025b0010ff2db6b646adb9c174571f367f27" gracePeriod=30 Nov 29 04:57:53 crc kubenswrapper[4799]: I1129 04:57:53.780622 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="33323158-ffcc-4f35-9797-e1fb18c3f0f1" containerName="ceilometer-notification-agent" containerID="cri-o://cd27c07c38df1fb6251e0b34585bb0d4585d48fb8bbeb278204694659344be9b" gracePeriod=30 Nov 29 04:57:53 crc kubenswrapper[4799]: I1129 04:57:53.788277 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="33323158-ffcc-4f35-9797-e1fb18c3f0f1" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.174:3000/\": read tcp 10.217.0.2:48424->10.217.0.174:3000: read: connection reset by peer" Nov 29 04:57:54 crc kubenswrapper[4799]: I1129 04:57:54.132359 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b856c5697-jgvwh" event={"ID":"c0b29191-5d29-44fa-9cde-c2604b5426bb","Type":"ContainerStarted","Data":"e02ae0768cfb16467fe1cee900bb77a9e6b8885860e9ce2e46625c69d51479b2"} Nov 29 04:57:54 crc kubenswrapper[4799]: I1129 04:57:54.132548 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b856c5697-jgvwh" Nov 29 04:57:54 crc kubenswrapper[4799]: I1129 04:57:54.137039 4799 generic.go:334] "Generic (PLEG): container finished" podID="33323158-ffcc-4f35-9797-e1fb18c3f0f1" containerID="ddabe9f210f448ee6eb45e8f616342f38b8ed24a13d6a5a5e213be02cd5fec35" exitCode=0 Nov 29 04:57:54 crc kubenswrapper[4799]: I1129 04:57:54.137078 4799 generic.go:334] "Generic (PLEG): container finished" podID="33323158-ffcc-4f35-9797-e1fb18c3f0f1" containerID="a2060acbc541160c4825d6d930f8025b0010ff2db6b646adb9c174571f367f27" exitCode=2 Nov 29 04:57:54 crc kubenswrapper[4799]: I1129 04:57:54.137340 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33323158-ffcc-4f35-9797-e1fb18c3f0f1","Type":"ContainerDied","Data":"ddabe9f210f448ee6eb45e8f616342f38b8ed24a13d6a5a5e213be02cd5fec35"} Nov 29 04:57:54 crc kubenswrapper[4799]: I1129 04:57:54.137379 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33323158-ffcc-4f35-9797-e1fb18c3f0f1","Type":"ContainerDied","Data":"a2060acbc541160c4825d6d930f8025b0010ff2db6b646adb9c174571f367f27"} Nov 29 04:57:54 crc kubenswrapper[4799]: I1129 04:57:54.151849 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b856c5697-jgvwh" podStartSLOduration=3.151826156 podStartE2EDuration="3.151826156s" podCreationTimestamp="2025-11-29 04:57:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:57:54.149569392 +0000 UTC m=+1149.792499822" watchObservedRunningTime="2025-11-29 04:57:54.151826156 +0000 UTC m=+1149.794756556" Nov 29 04:57:54 crc kubenswrapper[4799]: I1129 04:57:54.314897 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 29 04:57:54 crc kubenswrapper[4799]: I1129 04:57:54.315132 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="dea67977-de74-47f9-8db0-6d032d8579c2" containerName="nova-api-log" containerID="cri-o://3f364fedaf9c50e25e9c3ced9d715beb8e2933f1fff4b185d7648a9b2b4d8a44" gracePeriod=30 Nov 29 04:57:54 crc kubenswrapper[4799]: I1129 04:57:54.315465 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="dea67977-de74-47f9-8db0-6d032d8579c2" containerName="nova-api-api" containerID="cri-o://10c94e25337ea324516bbdcfe5f7c7d29e1e16dc5a878791e6ce18e3264e3dda" gracePeriod=30 Nov 29 04:57:55 crc kubenswrapper[4799]: I1129 04:57:55.150803 4799 generic.go:334] "Generic (PLEG): container finished" podID="33323158-ffcc-4f35-9797-e1fb18c3f0f1" containerID="9d0c07f09a3c0e24fb8bfd28d5f8c8ee44af9d249f5f1d1fb9b8f47a7f779370" exitCode=0 Nov 29 04:57:55 crc kubenswrapper[4799]: I1129 04:57:55.150846 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33323158-ffcc-4f35-9797-e1fb18c3f0f1","Type":"ContainerDied","Data":"9d0c07f09a3c0e24fb8bfd28d5f8c8ee44af9d249f5f1d1fb9b8f47a7f779370"} Nov 29 04:57:55 crc kubenswrapper[4799]: I1129 04:57:55.153317 4799 generic.go:334] "Generic (PLEG): container finished" podID="dea67977-de74-47f9-8db0-6d032d8579c2" containerID="3f364fedaf9c50e25e9c3ced9d715beb8e2933f1fff4b185d7648a9b2b4d8a44" exitCode=143 Nov 29 04:57:55 crc kubenswrapper[4799]: I1129 04:57:55.153442 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dea67977-de74-47f9-8db0-6d032d8579c2","Type":"ContainerDied","Data":"3f364fedaf9c50e25e9c3ced9d715beb8e2933f1fff4b185d7648a9b2b4d8a44"} Nov 29 04:57:56 crc kubenswrapper[4799]: I1129 04:57:56.517021 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 29 04:57:56 crc kubenswrapper[4799]: I1129 04:57:56.529596 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 29 04:57:56 crc kubenswrapper[4799]: I1129 04:57:56.529639 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 29 04:57:57 crc kubenswrapper[4799]: I1129 04:57:57.180523 4799 generic.go:334] "Generic (PLEG): container finished" podID="33323158-ffcc-4f35-9797-e1fb18c3f0f1" containerID="cd27c07c38df1fb6251e0b34585bb0d4585d48fb8bbeb278204694659344be9b" exitCode=0 Nov 29 04:57:57 crc kubenswrapper[4799]: I1129 04:57:57.180572 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33323158-ffcc-4f35-9797-e1fb18c3f0f1","Type":"ContainerDied","Data":"cd27c07c38df1fb6251e0b34585bb0d4585d48fb8bbeb278204694659344be9b"} Nov 29 04:57:57 crc kubenswrapper[4799]: I1129 04:57:57.591730 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 04:57:57 crc kubenswrapper[4799]: I1129 04:57:57.706588 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33323158-ffcc-4f35-9797-e1fb18c3f0f1-scripts\") pod \"33323158-ffcc-4f35-9797-e1fb18c3f0f1\" (UID: \"33323158-ffcc-4f35-9797-e1fb18c3f0f1\") " Nov 29 04:57:57 crc kubenswrapper[4799]: I1129 04:57:57.706666 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33323158-ffcc-4f35-9797-e1fb18c3f0f1-run-httpd\") pod \"33323158-ffcc-4f35-9797-e1fb18c3f0f1\" (UID: \"33323158-ffcc-4f35-9797-e1fb18c3f0f1\") " Nov 29 04:57:57 crc kubenswrapper[4799]: I1129 04:57:57.706697 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33323158-ffcc-4f35-9797-e1fb18c3f0f1-combined-ca-bundle\") pod \"33323158-ffcc-4f35-9797-e1fb18c3f0f1\" (UID: \"33323158-ffcc-4f35-9797-e1fb18c3f0f1\") " Nov 29 04:57:57 crc kubenswrapper[4799]: I1129 04:57:57.706773 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/33323158-ffcc-4f35-9797-e1fb18c3f0f1-sg-core-conf-yaml\") pod \"33323158-ffcc-4f35-9797-e1fb18c3f0f1\" (UID: \"33323158-ffcc-4f35-9797-e1fb18c3f0f1\") " Nov 29 04:57:57 crc kubenswrapper[4799]: I1129 04:57:57.706896 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/33323158-ffcc-4f35-9797-e1fb18c3f0f1-ceilometer-tls-certs\") pod \"33323158-ffcc-4f35-9797-e1fb18c3f0f1\" (UID: \"33323158-ffcc-4f35-9797-e1fb18c3f0f1\") " Nov 29 04:57:57 crc kubenswrapper[4799]: I1129 04:57:57.707009 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33323158-ffcc-4f35-9797-e1fb18c3f0f1-config-data\") pod \"33323158-ffcc-4f35-9797-e1fb18c3f0f1\" (UID: \"33323158-ffcc-4f35-9797-e1fb18c3f0f1\") " Nov 29 04:57:57 crc kubenswrapper[4799]: I1129 04:57:57.707092 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zghtt\" (UniqueName: \"kubernetes.io/projected/33323158-ffcc-4f35-9797-e1fb18c3f0f1-kube-api-access-zghtt\") pod \"33323158-ffcc-4f35-9797-e1fb18c3f0f1\" (UID: \"33323158-ffcc-4f35-9797-e1fb18c3f0f1\") " Nov 29 04:57:57 crc kubenswrapper[4799]: I1129 04:57:57.707176 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33323158-ffcc-4f35-9797-e1fb18c3f0f1-log-httpd\") pod \"33323158-ffcc-4f35-9797-e1fb18c3f0f1\" (UID: \"33323158-ffcc-4f35-9797-e1fb18c3f0f1\") " Nov 29 04:57:57 crc kubenswrapper[4799]: I1129 04:57:57.707492 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33323158-ffcc-4f35-9797-e1fb18c3f0f1-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "33323158-ffcc-4f35-9797-e1fb18c3f0f1" (UID: "33323158-ffcc-4f35-9797-e1fb18c3f0f1"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:57:57 crc kubenswrapper[4799]: I1129 04:57:57.707780 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33323158-ffcc-4f35-9797-e1fb18c3f0f1-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "33323158-ffcc-4f35-9797-e1fb18c3f0f1" (UID: "33323158-ffcc-4f35-9797-e1fb18c3f0f1"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:57:57 crc kubenswrapper[4799]: I1129 04:57:57.708005 4799 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33323158-ffcc-4f35-9797-e1fb18c3f0f1-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:57 crc kubenswrapper[4799]: I1129 04:57:57.708032 4799 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33323158-ffcc-4f35-9797-e1fb18c3f0f1-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:57 crc kubenswrapper[4799]: I1129 04:57:57.712838 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33323158-ffcc-4f35-9797-e1fb18c3f0f1-kube-api-access-zghtt" (OuterVolumeSpecName: "kube-api-access-zghtt") pod "33323158-ffcc-4f35-9797-e1fb18c3f0f1" (UID: "33323158-ffcc-4f35-9797-e1fb18c3f0f1"). InnerVolumeSpecName "kube-api-access-zghtt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:57:57 crc kubenswrapper[4799]: I1129 04:57:57.712982 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33323158-ffcc-4f35-9797-e1fb18c3f0f1-scripts" (OuterVolumeSpecName: "scripts") pod "33323158-ffcc-4f35-9797-e1fb18c3f0f1" (UID: "33323158-ffcc-4f35-9797-e1fb18c3f0f1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:57:57 crc kubenswrapper[4799]: I1129 04:57:57.754006 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33323158-ffcc-4f35-9797-e1fb18c3f0f1-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "33323158-ffcc-4f35-9797-e1fb18c3f0f1" (UID: "33323158-ffcc-4f35-9797-e1fb18c3f0f1"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:57:57 crc kubenswrapper[4799]: I1129 04:57:57.759665 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33323158-ffcc-4f35-9797-e1fb18c3f0f1-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "33323158-ffcc-4f35-9797-e1fb18c3f0f1" (UID: "33323158-ffcc-4f35-9797-e1fb18c3f0f1"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:57:57 crc kubenswrapper[4799]: I1129 04:57:57.774776 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33323158-ffcc-4f35-9797-e1fb18c3f0f1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "33323158-ffcc-4f35-9797-e1fb18c3f0f1" (UID: "33323158-ffcc-4f35-9797-e1fb18c3f0f1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:57:57 crc kubenswrapper[4799]: I1129 04:57:57.804540 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33323158-ffcc-4f35-9797-e1fb18c3f0f1-config-data" (OuterVolumeSpecName: "config-data") pod "33323158-ffcc-4f35-9797-e1fb18c3f0f1" (UID: "33323158-ffcc-4f35-9797-e1fb18c3f0f1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:57:57 crc kubenswrapper[4799]: I1129 04:57:57.809623 4799 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/33323158-ffcc-4f35-9797-e1fb18c3f0f1-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:57 crc kubenswrapper[4799]: I1129 04:57:57.809649 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33323158-ffcc-4f35-9797-e1fb18c3f0f1-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:57 crc kubenswrapper[4799]: I1129 04:57:57.809659 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zghtt\" (UniqueName: \"kubernetes.io/projected/33323158-ffcc-4f35-9797-e1fb18c3f0f1-kube-api-access-zghtt\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:57 crc kubenswrapper[4799]: I1129 04:57:57.809670 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33323158-ffcc-4f35-9797-e1fb18c3f0f1-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:57 crc kubenswrapper[4799]: I1129 04:57:57.809680 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33323158-ffcc-4f35-9797-e1fb18c3f0f1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:57 crc kubenswrapper[4799]: I1129 04:57:57.809691 4799 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/33323158-ffcc-4f35-9797-e1fb18c3f0f1-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.139332 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.193424 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33323158-ffcc-4f35-9797-e1fb18c3f0f1","Type":"ContainerDied","Data":"0a7339bd71d1f8c83aaef56fd7f9e1c198ef00141545d1a6859fd487919b6ffb"} Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.193445 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.193511 4799 scope.go:117] "RemoveContainer" containerID="ddabe9f210f448ee6eb45e8f616342f38b8ed24a13d6a5a5e213be02cd5fec35" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.201922 4799 generic.go:334] "Generic (PLEG): container finished" podID="dea67977-de74-47f9-8db0-6d032d8579c2" containerID="10c94e25337ea324516bbdcfe5f7c7d29e1e16dc5a878791e6ce18e3264e3dda" exitCode=0 Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.201993 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dea67977-de74-47f9-8db0-6d032d8579c2","Type":"ContainerDied","Data":"10c94e25337ea324516bbdcfe5f7c7d29e1e16dc5a878791e6ce18e3264e3dda"} Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.202031 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dea67977-de74-47f9-8db0-6d032d8579c2","Type":"ContainerDied","Data":"5edc1de60295dab1abfc53e0c8bd1308163896f0d5bb8931c90bda47aae8c6db"} Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.202040 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.273689 4799 scope.go:117] "RemoveContainer" containerID="a2060acbc541160c4825d6d930f8025b0010ff2db6b646adb9c174571f367f27" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.289858 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.300882 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.323194 4799 scope.go:117] "RemoveContainer" containerID="cd27c07c38df1fb6251e0b34585bb0d4585d48fb8bbeb278204694659344be9b" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.332303 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dea67977-de74-47f9-8db0-6d032d8579c2-logs\") pod \"dea67977-de74-47f9-8db0-6d032d8579c2\" (UID: \"dea67977-de74-47f9-8db0-6d032d8579c2\") " Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.332363 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dea67977-de74-47f9-8db0-6d032d8579c2-combined-ca-bundle\") pod \"dea67977-de74-47f9-8db0-6d032d8579c2\" (UID: \"dea67977-de74-47f9-8db0-6d032d8579c2\") " Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.332533 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4msgw\" (UniqueName: \"kubernetes.io/projected/dea67977-de74-47f9-8db0-6d032d8579c2-kube-api-access-4msgw\") pod \"dea67977-de74-47f9-8db0-6d032d8579c2\" (UID: \"dea67977-de74-47f9-8db0-6d032d8579c2\") " Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.332579 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dea67977-de74-47f9-8db0-6d032d8579c2-config-data\") pod \"dea67977-de74-47f9-8db0-6d032d8579c2\" (UID: \"dea67977-de74-47f9-8db0-6d032d8579c2\") " Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.332779 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dea67977-de74-47f9-8db0-6d032d8579c2-logs" (OuterVolumeSpecName: "logs") pod "dea67977-de74-47f9-8db0-6d032d8579c2" (UID: "dea67977-de74-47f9-8db0-6d032d8579c2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.335582 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dea67977-de74-47f9-8db0-6d032d8579c2-logs\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.346929 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dea67977-de74-47f9-8db0-6d032d8579c2-kube-api-access-4msgw" (OuterVolumeSpecName: "kube-api-access-4msgw") pod "dea67977-de74-47f9-8db0-6d032d8579c2" (UID: "dea67977-de74-47f9-8db0-6d032d8579c2"). InnerVolumeSpecName "kube-api-access-4msgw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.363735 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 29 04:57:58 crc kubenswrapper[4799]: E1129 04:57:58.364333 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33323158-ffcc-4f35-9797-e1fb18c3f0f1" containerName="sg-core" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.364359 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="33323158-ffcc-4f35-9797-e1fb18c3f0f1" containerName="sg-core" Nov 29 04:57:58 crc kubenswrapper[4799]: E1129 04:57:58.364382 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dea67977-de74-47f9-8db0-6d032d8579c2" containerName="nova-api-log" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.364392 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="dea67977-de74-47f9-8db0-6d032d8579c2" containerName="nova-api-log" Nov 29 04:57:58 crc kubenswrapper[4799]: E1129 04:57:58.364408 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33323158-ffcc-4f35-9797-e1fb18c3f0f1" containerName="proxy-httpd" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.364418 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="33323158-ffcc-4f35-9797-e1fb18c3f0f1" containerName="proxy-httpd" Nov 29 04:57:58 crc kubenswrapper[4799]: E1129 04:57:58.364435 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dea67977-de74-47f9-8db0-6d032d8579c2" containerName="nova-api-api" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.364443 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="dea67977-de74-47f9-8db0-6d032d8579c2" containerName="nova-api-api" Nov 29 04:57:58 crc kubenswrapper[4799]: E1129 04:57:58.364474 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33323158-ffcc-4f35-9797-e1fb18c3f0f1" containerName="ceilometer-notification-agent" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.364482 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="33323158-ffcc-4f35-9797-e1fb18c3f0f1" containerName="ceilometer-notification-agent" Nov 29 04:57:58 crc kubenswrapper[4799]: E1129 04:57:58.364500 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33323158-ffcc-4f35-9797-e1fb18c3f0f1" containerName="ceilometer-central-agent" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.364510 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="33323158-ffcc-4f35-9797-e1fb18c3f0f1" containerName="ceilometer-central-agent" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.364728 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="dea67977-de74-47f9-8db0-6d032d8579c2" containerName="nova-api-log" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.364750 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="33323158-ffcc-4f35-9797-e1fb18c3f0f1" containerName="sg-core" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.364764 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="dea67977-de74-47f9-8db0-6d032d8579c2" containerName="nova-api-api" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.364782 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="33323158-ffcc-4f35-9797-e1fb18c3f0f1" containerName="ceilometer-central-agent" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.365874 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="33323158-ffcc-4f35-9797-e1fb18c3f0f1" containerName="proxy-httpd" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.365889 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="33323158-ffcc-4f35-9797-e1fb18c3f0f1" containerName="ceilometer-notification-agent" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.368165 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.372387 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.372511 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.372761 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dea67977-de74-47f9-8db0-6d032d8579c2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dea67977-de74-47f9-8db0-6d032d8579c2" (UID: "dea67977-de74-47f9-8db0-6d032d8579c2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.376173 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.376633 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.380780 4799 scope.go:117] "RemoveContainer" containerID="9d0c07f09a3c0e24fb8bfd28d5f8c8ee44af9d249f5f1d1fb9b8f47a7f779370" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.422213 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dea67977-de74-47f9-8db0-6d032d8579c2-config-data" (OuterVolumeSpecName: "config-data") pod "dea67977-de74-47f9-8db0-6d032d8579c2" (UID: "dea67977-de74-47f9-8db0-6d032d8579c2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.427065 4799 scope.go:117] "RemoveContainer" containerID="10c94e25337ea324516bbdcfe5f7c7d29e1e16dc5a878791e6ce18e3264e3dda" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.439481 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dea67977-de74-47f9-8db0-6d032d8579c2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.439525 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4msgw\" (UniqueName: \"kubernetes.io/projected/dea67977-de74-47f9-8db0-6d032d8579c2-kube-api-access-4msgw\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.439539 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dea67977-de74-47f9-8db0-6d032d8579c2-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.455338 4799 scope.go:117] "RemoveContainer" containerID="3f364fedaf9c50e25e9c3ced9d715beb8e2933f1fff4b185d7648a9b2b4d8a44" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.477333 4799 scope.go:117] "RemoveContainer" containerID="10c94e25337ea324516bbdcfe5f7c7d29e1e16dc5a878791e6ce18e3264e3dda" Nov 29 04:57:58 crc kubenswrapper[4799]: E1129 04:57:58.477957 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10c94e25337ea324516bbdcfe5f7c7d29e1e16dc5a878791e6ce18e3264e3dda\": container with ID starting with 10c94e25337ea324516bbdcfe5f7c7d29e1e16dc5a878791e6ce18e3264e3dda not found: ID does not exist" containerID="10c94e25337ea324516bbdcfe5f7c7d29e1e16dc5a878791e6ce18e3264e3dda" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.478081 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10c94e25337ea324516bbdcfe5f7c7d29e1e16dc5a878791e6ce18e3264e3dda"} err="failed to get container status \"10c94e25337ea324516bbdcfe5f7c7d29e1e16dc5a878791e6ce18e3264e3dda\": rpc error: code = NotFound desc = could not find container \"10c94e25337ea324516bbdcfe5f7c7d29e1e16dc5a878791e6ce18e3264e3dda\": container with ID starting with 10c94e25337ea324516bbdcfe5f7c7d29e1e16dc5a878791e6ce18e3264e3dda not found: ID does not exist" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.478182 4799 scope.go:117] "RemoveContainer" containerID="3f364fedaf9c50e25e9c3ced9d715beb8e2933f1fff4b185d7648a9b2b4d8a44" Nov 29 04:57:58 crc kubenswrapper[4799]: E1129 04:57:58.478523 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f364fedaf9c50e25e9c3ced9d715beb8e2933f1fff4b185d7648a9b2b4d8a44\": container with ID starting with 3f364fedaf9c50e25e9c3ced9d715beb8e2933f1fff4b185d7648a9b2b4d8a44 not found: ID does not exist" containerID="3f364fedaf9c50e25e9c3ced9d715beb8e2933f1fff4b185d7648a9b2b4d8a44" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.478617 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f364fedaf9c50e25e9c3ced9d715beb8e2933f1fff4b185d7648a9b2b4d8a44"} err="failed to get container status \"3f364fedaf9c50e25e9c3ced9d715beb8e2933f1fff4b185d7648a9b2b4d8a44\": rpc error: code = NotFound desc = could not find container \"3f364fedaf9c50e25e9c3ced9d715beb8e2933f1fff4b185d7648a9b2b4d8a44\": container with ID starting with 3f364fedaf9c50e25e9c3ced9d715beb8e2933f1fff4b185d7648a9b2b4d8a44 not found: ID does not exist" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.541140 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e378b38-1cb5-4f77-b9a4-e6cfce8da188-config-data\") pod \"ceilometer-0\" (UID: \"9e378b38-1cb5-4f77-b9a4-e6cfce8da188\") " pod="openstack/ceilometer-0" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.541167 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.541217 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e378b38-1cb5-4f77-b9a4-e6cfce8da188-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9e378b38-1cb5-4f77-b9a4-e6cfce8da188\") " pod="openstack/ceilometer-0" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.541295 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9e378b38-1cb5-4f77-b9a4-e6cfce8da188-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9e378b38-1cb5-4f77-b9a4-e6cfce8da188\") " pod="openstack/ceilometer-0" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.541324 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9e378b38-1cb5-4f77-b9a4-e6cfce8da188-run-httpd\") pod \"ceilometer-0\" (UID: \"9e378b38-1cb5-4f77-b9a4-e6cfce8da188\") " pod="openstack/ceilometer-0" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.541343 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e378b38-1cb5-4f77-b9a4-e6cfce8da188-scripts\") pod \"ceilometer-0\" (UID: \"9e378b38-1cb5-4f77-b9a4-e6cfce8da188\") " pod="openstack/ceilometer-0" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.541370 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e378b38-1cb5-4f77-b9a4-e6cfce8da188-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9e378b38-1cb5-4f77-b9a4-e6cfce8da188\") " pod="openstack/ceilometer-0" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.541422 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9e378b38-1cb5-4f77-b9a4-e6cfce8da188-log-httpd\") pod \"ceilometer-0\" (UID: \"9e378b38-1cb5-4f77-b9a4-e6cfce8da188\") " pod="openstack/ceilometer-0" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.541496 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cfg5\" (UniqueName: \"kubernetes.io/projected/9e378b38-1cb5-4f77-b9a4-e6cfce8da188-kube-api-access-9cfg5\") pod \"ceilometer-0\" (UID: \"9e378b38-1cb5-4f77-b9a4-e6cfce8da188\") " pod="openstack/ceilometer-0" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.552663 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.563826 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.565694 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.568161 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.569355 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.570143 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.590407 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.642851 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e378b38-1cb5-4f77-b9a4-e6cfce8da188-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9e378b38-1cb5-4f77-b9a4-e6cfce8da188\") " pod="openstack/ceilometer-0" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.642946 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9e378b38-1cb5-4f77-b9a4-e6cfce8da188-log-httpd\") pod \"ceilometer-0\" (UID: \"9e378b38-1cb5-4f77-b9a4-e6cfce8da188\") " pod="openstack/ceilometer-0" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.643018 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9cfg5\" (UniqueName: \"kubernetes.io/projected/9e378b38-1cb5-4f77-b9a4-e6cfce8da188-kube-api-access-9cfg5\") pod \"ceilometer-0\" (UID: \"9e378b38-1cb5-4f77-b9a4-e6cfce8da188\") " pod="openstack/ceilometer-0" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.643056 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e378b38-1cb5-4f77-b9a4-e6cfce8da188-config-data\") pod \"ceilometer-0\" (UID: \"9e378b38-1cb5-4f77-b9a4-e6cfce8da188\") " pod="openstack/ceilometer-0" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.643085 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e378b38-1cb5-4f77-b9a4-e6cfce8da188-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9e378b38-1cb5-4f77-b9a4-e6cfce8da188\") " pod="openstack/ceilometer-0" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.643110 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9e378b38-1cb5-4f77-b9a4-e6cfce8da188-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9e378b38-1cb5-4f77-b9a4-e6cfce8da188\") " pod="openstack/ceilometer-0" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.643137 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9e378b38-1cb5-4f77-b9a4-e6cfce8da188-run-httpd\") pod \"ceilometer-0\" (UID: \"9e378b38-1cb5-4f77-b9a4-e6cfce8da188\") " pod="openstack/ceilometer-0" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.643158 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e378b38-1cb5-4f77-b9a4-e6cfce8da188-scripts\") pod \"ceilometer-0\" (UID: \"9e378b38-1cb5-4f77-b9a4-e6cfce8da188\") " pod="openstack/ceilometer-0" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.643559 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9e378b38-1cb5-4f77-b9a4-e6cfce8da188-log-httpd\") pod \"ceilometer-0\" (UID: \"9e378b38-1cb5-4f77-b9a4-e6cfce8da188\") " pod="openstack/ceilometer-0" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.644186 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9e378b38-1cb5-4f77-b9a4-e6cfce8da188-run-httpd\") pod \"ceilometer-0\" (UID: \"9e378b38-1cb5-4f77-b9a4-e6cfce8da188\") " pod="openstack/ceilometer-0" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.648046 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9e378b38-1cb5-4f77-b9a4-e6cfce8da188-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9e378b38-1cb5-4f77-b9a4-e6cfce8da188\") " pod="openstack/ceilometer-0" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.649104 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e378b38-1cb5-4f77-b9a4-e6cfce8da188-scripts\") pod \"ceilometer-0\" (UID: \"9e378b38-1cb5-4f77-b9a4-e6cfce8da188\") " pod="openstack/ceilometer-0" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.649187 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e378b38-1cb5-4f77-b9a4-e6cfce8da188-config-data\") pod \"ceilometer-0\" (UID: \"9e378b38-1cb5-4f77-b9a4-e6cfce8da188\") " pod="openstack/ceilometer-0" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.649466 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e378b38-1cb5-4f77-b9a4-e6cfce8da188-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9e378b38-1cb5-4f77-b9a4-e6cfce8da188\") " pod="openstack/ceilometer-0" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.649622 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e378b38-1cb5-4f77-b9a4-e6cfce8da188-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9e378b38-1cb5-4f77-b9a4-e6cfce8da188\") " pod="openstack/ceilometer-0" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.661111 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9cfg5\" (UniqueName: \"kubernetes.io/projected/9e378b38-1cb5-4f77-b9a4-e6cfce8da188-kube-api-access-9cfg5\") pod \"ceilometer-0\" (UID: \"9e378b38-1cb5-4f77-b9a4-e6cfce8da188\") " pod="openstack/ceilometer-0" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.672479 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33323158-ffcc-4f35-9797-e1fb18c3f0f1" path="/var/lib/kubelet/pods/33323158-ffcc-4f35-9797-e1fb18c3f0f1/volumes" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.673544 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dea67977-de74-47f9-8db0-6d032d8579c2" path="/var/lib/kubelet/pods/dea67977-de74-47f9-8db0-6d032d8579c2/volumes" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.708017 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.745602 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26f4f01b-2064-4b8c-896e-9cf708c47ebf-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"26f4f01b-2064-4b8c-896e-9cf708c47ebf\") " pod="openstack/nova-api-0" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.745676 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/26f4f01b-2064-4b8c-896e-9cf708c47ebf-public-tls-certs\") pod \"nova-api-0\" (UID: \"26f4f01b-2064-4b8c-896e-9cf708c47ebf\") " pod="openstack/nova-api-0" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.745835 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/26f4f01b-2064-4b8c-896e-9cf708c47ebf-internal-tls-certs\") pod \"nova-api-0\" (UID: \"26f4f01b-2064-4b8c-896e-9cf708c47ebf\") " pod="openstack/nova-api-0" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.745875 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26f4f01b-2064-4b8c-896e-9cf708c47ebf-config-data\") pod \"nova-api-0\" (UID: \"26f4f01b-2064-4b8c-896e-9cf708c47ebf\") " pod="openstack/nova-api-0" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.745944 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26f4f01b-2064-4b8c-896e-9cf708c47ebf-logs\") pod \"nova-api-0\" (UID: \"26f4f01b-2064-4b8c-896e-9cf708c47ebf\") " pod="openstack/nova-api-0" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.745981 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqt5f\" (UniqueName: \"kubernetes.io/projected/26f4f01b-2064-4b8c-896e-9cf708c47ebf-kube-api-access-dqt5f\") pod \"nova-api-0\" (UID: \"26f4f01b-2064-4b8c-896e-9cf708c47ebf\") " pod="openstack/nova-api-0" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.848366 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/26f4f01b-2064-4b8c-896e-9cf708c47ebf-internal-tls-certs\") pod \"nova-api-0\" (UID: \"26f4f01b-2064-4b8c-896e-9cf708c47ebf\") " pod="openstack/nova-api-0" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.848425 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26f4f01b-2064-4b8c-896e-9cf708c47ebf-config-data\") pod \"nova-api-0\" (UID: \"26f4f01b-2064-4b8c-896e-9cf708c47ebf\") " pod="openstack/nova-api-0" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.848521 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26f4f01b-2064-4b8c-896e-9cf708c47ebf-logs\") pod \"nova-api-0\" (UID: \"26f4f01b-2064-4b8c-896e-9cf708c47ebf\") " pod="openstack/nova-api-0" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.848560 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqt5f\" (UniqueName: \"kubernetes.io/projected/26f4f01b-2064-4b8c-896e-9cf708c47ebf-kube-api-access-dqt5f\") pod \"nova-api-0\" (UID: \"26f4f01b-2064-4b8c-896e-9cf708c47ebf\") " pod="openstack/nova-api-0" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.848633 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26f4f01b-2064-4b8c-896e-9cf708c47ebf-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"26f4f01b-2064-4b8c-896e-9cf708c47ebf\") " pod="openstack/nova-api-0" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.848669 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/26f4f01b-2064-4b8c-896e-9cf708c47ebf-public-tls-certs\") pod \"nova-api-0\" (UID: \"26f4f01b-2064-4b8c-896e-9cf708c47ebf\") " pod="openstack/nova-api-0" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.849311 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26f4f01b-2064-4b8c-896e-9cf708c47ebf-logs\") pod \"nova-api-0\" (UID: \"26f4f01b-2064-4b8c-896e-9cf708c47ebf\") " pod="openstack/nova-api-0" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.854375 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/26f4f01b-2064-4b8c-896e-9cf708c47ebf-internal-tls-certs\") pod \"nova-api-0\" (UID: \"26f4f01b-2064-4b8c-896e-9cf708c47ebf\") " pod="openstack/nova-api-0" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.854782 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26f4f01b-2064-4b8c-896e-9cf708c47ebf-config-data\") pod \"nova-api-0\" (UID: \"26f4f01b-2064-4b8c-896e-9cf708c47ebf\") " pod="openstack/nova-api-0" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.855699 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26f4f01b-2064-4b8c-896e-9cf708c47ebf-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"26f4f01b-2064-4b8c-896e-9cf708c47ebf\") " pod="openstack/nova-api-0" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.861565 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/26f4f01b-2064-4b8c-896e-9cf708c47ebf-public-tls-certs\") pod \"nova-api-0\" (UID: \"26f4f01b-2064-4b8c-896e-9cf708c47ebf\") " pod="openstack/nova-api-0" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.872663 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqt5f\" (UniqueName: \"kubernetes.io/projected/26f4f01b-2064-4b8c-896e-9cf708c47ebf-kube-api-access-dqt5f\") pod \"nova-api-0\" (UID: \"26f4f01b-2064-4b8c-896e-9cf708c47ebf\") " pod="openstack/nova-api-0" Nov 29 04:57:58 crc kubenswrapper[4799]: I1129 04:57:58.897922 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 29 04:57:59 crc kubenswrapper[4799]: I1129 04:57:59.206504 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 29 04:57:59 crc kubenswrapper[4799]: I1129 04:57:59.381182 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 29 04:57:59 crc kubenswrapper[4799]: W1129 04:57:59.383668 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod26f4f01b_2064_4b8c_896e_9cf708c47ebf.slice/crio-7e144b288b33cf1c9a2bbcbae169de58b99486ee172de18e655c1771cda8a649 WatchSource:0}: Error finding container 7e144b288b33cf1c9a2bbcbae169de58b99486ee172de18e655c1771cda8a649: Status 404 returned error can't find the container with id 7e144b288b33cf1c9a2bbcbae169de58b99486ee172de18e655c1771cda8a649 Nov 29 04:58:00 crc kubenswrapper[4799]: I1129 04:58:00.226516 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9e378b38-1cb5-4f77-b9a4-e6cfce8da188","Type":"ContainerStarted","Data":"f72ac746d459003c055d5bd70bdc1d61bb792cd8aa649c5218ef2ca22d8f7cf7"} Nov 29 04:58:00 crc kubenswrapper[4799]: I1129 04:58:00.229522 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"26f4f01b-2064-4b8c-896e-9cf708c47ebf","Type":"ContainerStarted","Data":"1e85574ad1103ec2a404861fcfc60e48c1adda765ded997e03bd6f0da1dbd49e"} Nov 29 04:58:00 crc kubenswrapper[4799]: I1129 04:58:00.229550 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"26f4f01b-2064-4b8c-896e-9cf708c47ebf","Type":"ContainerStarted","Data":"dd097e8aed8ba6307b766d41135154ddeee364b896e797b1d53adbd27bae566d"} Nov 29 04:58:00 crc kubenswrapper[4799]: I1129 04:58:00.229561 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"26f4f01b-2064-4b8c-896e-9cf708c47ebf","Type":"ContainerStarted","Data":"7e144b288b33cf1c9a2bbcbae169de58b99486ee172de18e655c1771cda8a649"} Nov 29 04:58:00 crc kubenswrapper[4799]: I1129 04:58:00.262515 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.2624891639999998 podStartE2EDuration="2.262489164s" podCreationTimestamp="2025-11-29 04:57:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:58:00.251655596 +0000 UTC m=+1155.894585996" watchObservedRunningTime="2025-11-29 04:58:00.262489164 +0000 UTC m=+1155.905419564" Nov 29 04:58:01 crc kubenswrapper[4799]: I1129 04:58:01.245515 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9e378b38-1cb5-4f77-b9a4-e6cfce8da188","Type":"ContainerStarted","Data":"d09b5086c0cd8c7c5bb69705d115dec19ed06948186960f1aa1eafe1b7cd88a8"} Nov 29 04:58:01 crc kubenswrapper[4799]: I1129 04:58:01.245853 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9e378b38-1cb5-4f77-b9a4-e6cfce8da188","Type":"ContainerStarted","Data":"86e09030d07ac79226803571fd562d1b65201ff6597388493a2af883ba02cab8"} Nov 29 04:58:01 crc kubenswrapper[4799]: I1129 04:58:01.517319 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Nov 29 04:58:01 crc kubenswrapper[4799]: I1129 04:58:01.530324 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 29 04:58:01 crc kubenswrapper[4799]: I1129 04:58:01.530385 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 29 04:58:01 crc kubenswrapper[4799]: I1129 04:58:01.541835 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Nov 29 04:58:01 crc kubenswrapper[4799]: I1129 04:58:01.668017 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b856c5697-jgvwh" Nov 29 04:58:01 crc kubenswrapper[4799]: I1129 04:58:01.764409 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-566b5b7845-8frfb"] Nov 29 04:58:01 crc kubenswrapper[4799]: I1129 04:58:01.764783 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-566b5b7845-8frfb" podUID="00ffa619-b06e-46bc-832a-f7fbc2e0efff" containerName="dnsmasq-dns" containerID="cri-o://cd2b609ab12cb328932676754756fc40228b18f8b81ffb0ef43d61eb2664f941" gracePeriod=10 Nov 29 04:58:02 crc kubenswrapper[4799]: I1129 04:58:02.261339 4799 generic.go:334] "Generic (PLEG): container finished" podID="00ffa619-b06e-46bc-832a-f7fbc2e0efff" containerID="cd2b609ab12cb328932676754756fc40228b18f8b81ffb0ef43d61eb2664f941" exitCode=0 Nov 29 04:58:02 crc kubenswrapper[4799]: I1129 04:58:02.262817 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-566b5b7845-8frfb" event={"ID":"00ffa619-b06e-46bc-832a-f7fbc2e0efff","Type":"ContainerDied","Data":"cd2b609ab12cb328932676754756fc40228b18f8b81ffb0ef43d61eb2664f941"} Nov 29 04:58:02 crc kubenswrapper[4799]: I1129 04:58:02.262867 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-566b5b7845-8frfb" event={"ID":"00ffa619-b06e-46bc-832a-f7fbc2e0efff","Type":"ContainerDied","Data":"6697d3b3a1c48517b7807ce467aaca6e6447664573c9533119068982bbde9b55"} Nov 29 04:58:02 crc kubenswrapper[4799]: I1129 04:58:02.262883 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6697d3b3a1c48517b7807ce467aaca6e6447664573c9533119068982bbde9b55" Nov 29 04:58:02 crc kubenswrapper[4799]: I1129 04:58:02.267192 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9e378b38-1cb5-4f77-b9a4-e6cfce8da188","Type":"ContainerStarted","Data":"adc14bfee676568c6a1b21f35b94c262a6fe0699e51bf4453d9d71b62855b2d8"} Nov 29 04:58:02 crc kubenswrapper[4799]: I1129 04:58:02.282903 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-566b5b7845-8frfb" Nov 29 04:58:02 crc kubenswrapper[4799]: I1129 04:58:02.290502 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Nov 29 04:58:02 crc kubenswrapper[4799]: I1129 04:58:02.448463 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00ffa619-b06e-46bc-832a-f7fbc2e0efff-config\") pod \"00ffa619-b06e-46bc-832a-f7fbc2e0efff\" (UID: \"00ffa619-b06e-46bc-832a-f7fbc2e0efff\") " Nov 29 04:58:02 crc kubenswrapper[4799]: I1129 04:58:02.448577 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjv8g\" (UniqueName: \"kubernetes.io/projected/00ffa619-b06e-46bc-832a-f7fbc2e0efff-kube-api-access-wjv8g\") pod \"00ffa619-b06e-46bc-832a-f7fbc2e0efff\" (UID: \"00ffa619-b06e-46bc-832a-f7fbc2e0efff\") " Nov 29 04:58:02 crc kubenswrapper[4799]: I1129 04:58:02.448682 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00ffa619-b06e-46bc-832a-f7fbc2e0efff-dns-svc\") pod \"00ffa619-b06e-46bc-832a-f7fbc2e0efff\" (UID: \"00ffa619-b06e-46bc-832a-f7fbc2e0efff\") " Nov 29 04:58:02 crc kubenswrapper[4799]: I1129 04:58:02.448769 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/00ffa619-b06e-46bc-832a-f7fbc2e0efff-ovsdbserver-sb\") pod \"00ffa619-b06e-46bc-832a-f7fbc2e0efff\" (UID: \"00ffa619-b06e-46bc-832a-f7fbc2e0efff\") " Nov 29 04:58:02 crc kubenswrapper[4799]: I1129 04:58:02.448853 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/00ffa619-b06e-46bc-832a-f7fbc2e0efff-ovsdbserver-nb\") pod \"00ffa619-b06e-46bc-832a-f7fbc2e0efff\" (UID: \"00ffa619-b06e-46bc-832a-f7fbc2e0efff\") " Nov 29 04:58:02 crc kubenswrapper[4799]: I1129 04:58:02.461808 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00ffa619-b06e-46bc-832a-f7fbc2e0efff-kube-api-access-wjv8g" (OuterVolumeSpecName: "kube-api-access-wjv8g") pod "00ffa619-b06e-46bc-832a-f7fbc2e0efff" (UID: "00ffa619-b06e-46bc-832a-f7fbc2e0efff"). InnerVolumeSpecName "kube-api-access-wjv8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:58:02 crc kubenswrapper[4799]: I1129 04:58:02.504805 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00ffa619-b06e-46bc-832a-f7fbc2e0efff-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "00ffa619-b06e-46bc-832a-f7fbc2e0efff" (UID: "00ffa619-b06e-46bc-832a-f7fbc2e0efff"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:58:02 crc kubenswrapper[4799]: I1129 04:58:02.510581 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-m5bjj"] Nov 29 04:58:02 crc kubenswrapper[4799]: E1129 04:58:02.511063 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00ffa619-b06e-46bc-832a-f7fbc2e0efff" containerName="init" Nov 29 04:58:02 crc kubenswrapper[4799]: I1129 04:58:02.511083 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="00ffa619-b06e-46bc-832a-f7fbc2e0efff" containerName="init" Nov 29 04:58:02 crc kubenswrapper[4799]: E1129 04:58:02.511117 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00ffa619-b06e-46bc-832a-f7fbc2e0efff" containerName="dnsmasq-dns" Nov 29 04:58:02 crc kubenswrapper[4799]: I1129 04:58:02.511125 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="00ffa619-b06e-46bc-832a-f7fbc2e0efff" containerName="dnsmasq-dns" Nov 29 04:58:02 crc kubenswrapper[4799]: I1129 04:58:02.511321 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="00ffa619-b06e-46bc-832a-f7fbc2e0efff" containerName="dnsmasq-dns" Nov 29 04:58:02 crc kubenswrapper[4799]: I1129 04:58:02.512993 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-m5bjj" Nov 29 04:58:02 crc kubenswrapper[4799]: I1129 04:58:02.521996 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Nov 29 04:58:02 crc kubenswrapper[4799]: I1129 04:58:02.522257 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Nov 29 04:58:02 crc kubenswrapper[4799]: I1129 04:58:02.529000 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00ffa619-b06e-46bc-832a-f7fbc2e0efff-config" (OuterVolumeSpecName: "config") pod "00ffa619-b06e-46bc-832a-f7fbc2e0efff" (UID: "00ffa619-b06e-46bc-832a-f7fbc2e0efff"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:58:02 crc kubenswrapper[4799]: I1129 04:58:02.539463 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-m5bjj"] Nov 29 04:58:02 crc kubenswrapper[4799]: I1129 04:58:02.547275 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00ffa619-b06e-46bc-832a-f7fbc2e0efff-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "00ffa619-b06e-46bc-832a-f7fbc2e0efff" (UID: "00ffa619-b06e-46bc-832a-f7fbc2e0efff"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:58:02 crc kubenswrapper[4799]: I1129 04:58:02.550648 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/00ffa619-b06e-46bc-832a-f7fbc2e0efff-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 29 04:58:02 crc kubenswrapper[4799]: I1129 04:58:02.550680 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/00ffa619-b06e-46bc-832a-f7fbc2e0efff-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 29 04:58:02 crc kubenswrapper[4799]: I1129 04:58:02.550693 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00ffa619-b06e-46bc-832a-f7fbc2e0efff-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:58:02 crc kubenswrapper[4799]: I1129 04:58:02.550704 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjv8g\" (UniqueName: \"kubernetes.io/projected/00ffa619-b06e-46bc-832a-f7fbc2e0efff-kube-api-access-wjv8g\") on node \"crc\" DevicePath \"\"" Nov 29 04:58:02 crc kubenswrapper[4799]: I1129 04:58:02.559825 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00ffa619-b06e-46bc-832a-f7fbc2e0efff-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "00ffa619-b06e-46bc-832a-f7fbc2e0efff" (UID: "00ffa619-b06e-46bc-832a-f7fbc2e0efff"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:58:02 crc kubenswrapper[4799]: I1129 04:58:02.583165 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="4e2d8276-3c6c-418a-8d82-7b574dc1dd63" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.178:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 29 04:58:02 crc kubenswrapper[4799]: I1129 04:58:02.583831 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="4e2d8276-3c6c-418a-8d82-7b574dc1dd63" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.178:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 29 04:58:02 crc kubenswrapper[4799]: I1129 04:58:02.652332 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dt4v\" (UniqueName: \"kubernetes.io/projected/35dd3ae0-d97c-49ff-8b70-a394fd58af0f-kube-api-access-4dt4v\") pod \"nova-cell1-cell-mapping-m5bjj\" (UID: \"35dd3ae0-d97c-49ff-8b70-a394fd58af0f\") " pod="openstack/nova-cell1-cell-mapping-m5bjj" Nov 29 04:58:02 crc kubenswrapper[4799]: I1129 04:58:02.652716 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35dd3ae0-d97c-49ff-8b70-a394fd58af0f-scripts\") pod \"nova-cell1-cell-mapping-m5bjj\" (UID: \"35dd3ae0-d97c-49ff-8b70-a394fd58af0f\") " pod="openstack/nova-cell1-cell-mapping-m5bjj" Nov 29 04:58:02 crc kubenswrapper[4799]: I1129 04:58:02.652915 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35dd3ae0-d97c-49ff-8b70-a394fd58af0f-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-m5bjj\" (UID: \"35dd3ae0-d97c-49ff-8b70-a394fd58af0f\") " pod="openstack/nova-cell1-cell-mapping-m5bjj" Nov 29 04:58:02 crc kubenswrapper[4799]: I1129 04:58:02.653041 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35dd3ae0-d97c-49ff-8b70-a394fd58af0f-config-data\") pod \"nova-cell1-cell-mapping-m5bjj\" (UID: \"35dd3ae0-d97c-49ff-8b70-a394fd58af0f\") " pod="openstack/nova-cell1-cell-mapping-m5bjj" Nov 29 04:58:02 crc kubenswrapper[4799]: I1129 04:58:02.653256 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00ffa619-b06e-46bc-832a-f7fbc2e0efff-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 29 04:58:02 crc kubenswrapper[4799]: I1129 04:58:02.755685 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35dd3ae0-d97c-49ff-8b70-a394fd58af0f-scripts\") pod \"nova-cell1-cell-mapping-m5bjj\" (UID: \"35dd3ae0-d97c-49ff-8b70-a394fd58af0f\") " pod="openstack/nova-cell1-cell-mapping-m5bjj" Nov 29 04:58:02 crc kubenswrapper[4799]: I1129 04:58:02.755983 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35dd3ae0-d97c-49ff-8b70-a394fd58af0f-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-m5bjj\" (UID: \"35dd3ae0-d97c-49ff-8b70-a394fd58af0f\") " pod="openstack/nova-cell1-cell-mapping-m5bjj" Nov 29 04:58:02 crc kubenswrapper[4799]: I1129 04:58:02.756107 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35dd3ae0-d97c-49ff-8b70-a394fd58af0f-config-data\") pod \"nova-cell1-cell-mapping-m5bjj\" (UID: \"35dd3ae0-d97c-49ff-8b70-a394fd58af0f\") " pod="openstack/nova-cell1-cell-mapping-m5bjj" Nov 29 04:58:02 crc kubenswrapper[4799]: I1129 04:58:02.756457 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dt4v\" (UniqueName: \"kubernetes.io/projected/35dd3ae0-d97c-49ff-8b70-a394fd58af0f-kube-api-access-4dt4v\") pod \"nova-cell1-cell-mapping-m5bjj\" (UID: \"35dd3ae0-d97c-49ff-8b70-a394fd58af0f\") " pod="openstack/nova-cell1-cell-mapping-m5bjj" Nov 29 04:58:02 crc kubenswrapper[4799]: I1129 04:58:02.761553 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35dd3ae0-d97c-49ff-8b70-a394fd58af0f-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-m5bjj\" (UID: \"35dd3ae0-d97c-49ff-8b70-a394fd58af0f\") " pod="openstack/nova-cell1-cell-mapping-m5bjj" Nov 29 04:58:02 crc kubenswrapper[4799]: I1129 04:58:02.761738 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35dd3ae0-d97c-49ff-8b70-a394fd58af0f-scripts\") pod \"nova-cell1-cell-mapping-m5bjj\" (UID: \"35dd3ae0-d97c-49ff-8b70-a394fd58af0f\") " pod="openstack/nova-cell1-cell-mapping-m5bjj" Nov 29 04:58:02 crc kubenswrapper[4799]: I1129 04:58:02.762045 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35dd3ae0-d97c-49ff-8b70-a394fd58af0f-config-data\") pod \"nova-cell1-cell-mapping-m5bjj\" (UID: \"35dd3ae0-d97c-49ff-8b70-a394fd58af0f\") " pod="openstack/nova-cell1-cell-mapping-m5bjj" Nov 29 04:58:02 crc kubenswrapper[4799]: I1129 04:58:02.781532 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dt4v\" (UniqueName: \"kubernetes.io/projected/35dd3ae0-d97c-49ff-8b70-a394fd58af0f-kube-api-access-4dt4v\") pod \"nova-cell1-cell-mapping-m5bjj\" (UID: \"35dd3ae0-d97c-49ff-8b70-a394fd58af0f\") " pod="openstack/nova-cell1-cell-mapping-m5bjj" Nov 29 04:58:02 crc kubenswrapper[4799]: I1129 04:58:02.847736 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-m5bjj" Nov 29 04:58:03 crc kubenswrapper[4799]: I1129 04:58:03.273096 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-566b5b7845-8frfb" Nov 29 04:58:03 crc kubenswrapper[4799]: I1129 04:58:03.299835 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-566b5b7845-8frfb"] Nov 29 04:58:03 crc kubenswrapper[4799]: I1129 04:58:03.307752 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-566b5b7845-8frfb"] Nov 29 04:58:03 crc kubenswrapper[4799]: I1129 04:58:03.330345 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-m5bjj"] Nov 29 04:58:04 crc kubenswrapper[4799]: I1129 04:58:04.285409 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9e378b38-1cb5-4f77-b9a4-e6cfce8da188","Type":"ContainerStarted","Data":"be4294eae32e4ed3aef4ebf11a13fbb8fb8da2c722a9d820a58674f8e9a082dc"} Nov 29 04:58:04 crc kubenswrapper[4799]: I1129 04:58:04.286120 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 29 04:58:04 crc kubenswrapper[4799]: I1129 04:58:04.288441 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-m5bjj" event={"ID":"35dd3ae0-d97c-49ff-8b70-a394fd58af0f","Type":"ContainerStarted","Data":"e1ec5147edd224f00432287e8890d08d4f3a9e3e7e0790d3533c7706df9da228"} Nov 29 04:58:04 crc kubenswrapper[4799]: I1129 04:58:04.288468 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-m5bjj" event={"ID":"35dd3ae0-d97c-49ff-8b70-a394fd58af0f","Type":"ContainerStarted","Data":"258a7e34a6be837a63d8d275cf4024cba6b5b50584c884004398e53e9484a546"} Nov 29 04:58:04 crc kubenswrapper[4799]: I1129 04:58:04.322395 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.120862403 podStartE2EDuration="6.322363139s" podCreationTimestamp="2025-11-29 04:57:58 +0000 UTC" firstStartedPulling="2025-11-29 04:57:59.214423862 +0000 UTC m=+1154.857354262" lastFinishedPulling="2025-11-29 04:58:03.415924598 +0000 UTC m=+1159.058854998" observedRunningTime="2025-11-29 04:58:04.310028476 +0000 UTC m=+1159.952958886" watchObservedRunningTime="2025-11-29 04:58:04.322363139 +0000 UTC m=+1159.965293539" Nov 29 04:58:04 crc kubenswrapper[4799]: I1129 04:58:04.340576 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-m5bjj" podStartSLOduration=2.340548441 podStartE2EDuration="2.340548441s" podCreationTimestamp="2025-11-29 04:58:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:58:04.333855273 +0000 UTC m=+1159.976785683" watchObservedRunningTime="2025-11-29 04:58:04.340548441 +0000 UTC m=+1159.983478841" Nov 29 04:58:04 crc kubenswrapper[4799]: I1129 04:58:04.670217 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00ffa619-b06e-46bc-832a-f7fbc2e0efff" path="/var/lib/kubelet/pods/00ffa619-b06e-46bc-832a-f7fbc2e0efff/volumes" Nov 29 04:58:07 crc kubenswrapper[4799]: I1129 04:58:07.498261 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 04:58:07 crc kubenswrapper[4799]: I1129 04:58:07.499184 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 04:58:08 crc kubenswrapper[4799]: I1129 04:58:08.901821 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 29 04:58:08 crc kubenswrapper[4799]: I1129 04:58:08.902573 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 29 04:58:09 crc kubenswrapper[4799]: I1129 04:58:09.335774 4799 generic.go:334] "Generic (PLEG): container finished" podID="35dd3ae0-d97c-49ff-8b70-a394fd58af0f" containerID="e1ec5147edd224f00432287e8890d08d4f3a9e3e7e0790d3533c7706df9da228" exitCode=0 Nov 29 04:58:09 crc kubenswrapper[4799]: I1129 04:58:09.335844 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-m5bjj" event={"ID":"35dd3ae0-d97c-49ff-8b70-a394fd58af0f","Type":"ContainerDied","Data":"e1ec5147edd224f00432287e8890d08d4f3a9e3e7e0790d3533c7706df9da228"} Nov 29 04:58:09 crc kubenswrapper[4799]: I1129 04:58:09.911097 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="26f4f01b-2064-4b8c-896e-9cf708c47ebf" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.181:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 29 04:58:09 crc kubenswrapper[4799]: I1129 04:58:09.911101 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="26f4f01b-2064-4b8c-896e-9cf708c47ebf" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.181:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 29 04:58:10 crc kubenswrapper[4799]: I1129 04:58:10.777235 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-m5bjj" Nov 29 04:58:10 crc kubenswrapper[4799]: I1129 04:58:10.934296 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35dd3ae0-d97c-49ff-8b70-a394fd58af0f-scripts\") pod \"35dd3ae0-d97c-49ff-8b70-a394fd58af0f\" (UID: \"35dd3ae0-d97c-49ff-8b70-a394fd58af0f\") " Nov 29 04:58:10 crc kubenswrapper[4799]: I1129 04:58:10.934450 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4dt4v\" (UniqueName: \"kubernetes.io/projected/35dd3ae0-d97c-49ff-8b70-a394fd58af0f-kube-api-access-4dt4v\") pod \"35dd3ae0-d97c-49ff-8b70-a394fd58af0f\" (UID: \"35dd3ae0-d97c-49ff-8b70-a394fd58af0f\") " Nov 29 04:58:10 crc kubenswrapper[4799]: I1129 04:58:10.934508 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35dd3ae0-d97c-49ff-8b70-a394fd58af0f-config-data\") pod \"35dd3ae0-d97c-49ff-8b70-a394fd58af0f\" (UID: \"35dd3ae0-d97c-49ff-8b70-a394fd58af0f\") " Nov 29 04:58:10 crc kubenswrapper[4799]: I1129 04:58:10.934529 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35dd3ae0-d97c-49ff-8b70-a394fd58af0f-combined-ca-bundle\") pod \"35dd3ae0-d97c-49ff-8b70-a394fd58af0f\" (UID: \"35dd3ae0-d97c-49ff-8b70-a394fd58af0f\") " Nov 29 04:58:10 crc kubenswrapper[4799]: I1129 04:58:10.941451 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35dd3ae0-d97c-49ff-8b70-a394fd58af0f-kube-api-access-4dt4v" (OuterVolumeSpecName: "kube-api-access-4dt4v") pod "35dd3ae0-d97c-49ff-8b70-a394fd58af0f" (UID: "35dd3ae0-d97c-49ff-8b70-a394fd58af0f"). InnerVolumeSpecName "kube-api-access-4dt4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:58:10 crc kubenswrapper[4799]: I1129 04:58:10.957367 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35dd3ae0-d97c-49ff-8b70-a394fd58af0f-scripts" (OuterVolumeSpecName: "scripts") pod "35dd3ae0-d97c-49ff-8b70-a394fd58af0f" (UID: "35dd3ae0-d97c-49ff-8b70-a394fd58af0f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:58:10 crc kubenswrapper[4799]: I1129 04:58:10.961822 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35dd3ae0-d97c-49ff-8b70-a394fd58af0f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "35dd3ae0-d97c-49ff-8b70-a394fd58af0f" (UID: "35dd3ae0-d97c-49ff-8b70-a394fd58af0f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:58:10 crc kubenswrapper[4799]: I1129 04:58:10.964411 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35dd3ae0-d97c-49ff-8b70-a394fd58af0f-config-data" (OuterVolumeSpecName: "config-data") pod "35dd3ae0-d97c-49ff-8b70-a394fd58af0f" (UID: "35dd3ae0-d97c-49ff-8b70-a394fd58af0f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:58:11 crc kubenswrapper[4799]: I1129 04:58:11.037436 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4dt4v\" (UniqueName: \"kubernetes.io/projected/35dd3ae0-d97c-49ff-8b70-a394fd58af0f-kube-api-access-4dt4v\") on node \"crc\" DevicePath \"\"" Nov 29 04:58:11 crc kubenswrapper[4799]: I1129 04:58:11.037484 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35dd3ae0-d97c-49ff-8b70-a394fd58af0f-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 04:58:11 crc kubenswrapper[4799]: I1129 04:58:11.037496 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35dd3ae0-d97c-49ff-8b70-a394fd58af0f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 04:58:11 crc kubenswrapper[4799]: I1129 04:58:11.037509 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35dd3ae0-d97c-49ff-8b70-a394fd58af0f-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 04:58:11 crc kubenswrapper[4799]: I1129 04:58:11.355673 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-m5bjj" event={"ID":"35dd3ae0-d97c-49ff-8b70-a394fd58af0f","Type":"ContainerDied","Data":"258a7e34a6be837a63d8d275cf4024cba6b5b50584c884004398e53e9484a546"} Nov 29 04:58:11 crc kubenswrapper[4799]: I1129 04:58:11.355737 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="258a7e34a6be837a63d8d275cf4024cba6b5b50584c884004398e53e9484a546" Nov 29 04:58:11 crc kubenswrapper[4799]: I1129 04:58:11.355907 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-m5bjj" Nov 29 04:58:11 crc kubenswrapper[4799]: I1129 04:58:11.564563 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 29 04:58:11 crc kubenswrapper[4799]: I1129 04:58:11.565921 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 29 04:58:11 crc kubenswrapper[4799]: I1129 04:58:11.573760 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 29 04:58:11 crc kubenswrapper[4799]: I1129 04:58:11.574168 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="26f4f01b-2064-4b8c-896e-9cf708c47ebf" containerName="nova-api-log" containerID="cri-o://dd097e8aed8ba6307b766d41135154ddeee364b896e797b1d53adbd27bae566d" gracePeriod=30 Nov 29 04:58:11 crc kubenswrapper[4799]: I1129 04:58:11.574267 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="26f4f01b-2064-4b8c-896e-9cf708c47ebf" containerName="nova-api-api" containerID="cri-o://1e85574ad1103ec2a404861fcfc60e48c1adda765ded997e03bd6f0da1dbd49e" gracePeriod=30 Nov 29 04:58:11 crc kubenswrapper[4799]: I1129 04:58:11.586732 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 29 04:58:11 crc kubenswrapper[4799]: I1129 04:58:11.588097 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="3c25c970-ef30-438b-b5d3-a98976fa7caa" containerName="nova-scheduler-scheduler" containerID="cri-o://1879026d9dc12823fb9939d498c22c08936c600c692b8f34e8b5385e53f855f4" gracePeriod=30 Nov 29 04:58:11 crc kubenswrapper[4799]: I1129 04:58:11.602148 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 29 04:58:11 crc kubenswrapper[4799]: I1129 04:58:11.612421 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 29 04:58:12 crc kubenswrapper[4799]: E1129 04:58:12.317558 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1879026d9dc12823fb9939d498c22c08936c600c692b8f34e8b5385e53f855f4" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 29 04:58:12 crc kubenswrapper[4799]: E1129 04:58:12.319352 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1879026d9dc12823fb9939d498c22c08936c600c692b8f34e8b5385e53f855f4" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 29 04:58:12 crc kubenswrapper[4799]: E1129 04:58:12.320942 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1879026d9dc12823fb9939d498c22c08936c600c692b8f34e8b5385e53f855f4" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 29 04:58:12 crc kubenswrapper[4799]: E1129 04:58:12.320984 4799 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="3c25c970-ef30-438b-b5d3-a98976fa7caa" containerName="nova-scheduler-scheduler" Nov 29 04:58:12 crc kubenswrapper[4799]: I1129 04:58:12.375727 4799 generic.go:334] "Generic (PLEG): container finished" podID="26f4f01b-2064-4b8c-896e-9cf708c47ebf" containerID="dd097e8aed8ba6307b766d41135154ddeee364b896e797b1d53adbd27bae566d" exitCode=143 Nov 29 04:58:12 crc kubenswrapper[4799]: I1129 04:58:12.377477 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"26f4f01b-2064-4b8c-896e-9cf708c47ebf","Type":"ContainerDied","Data":"dd097e8aed8ba6307b766d41135154ddeee364b896e797b1d53adbd27bae566d"} Nov 29 04:58:12 crc kubenswrapper[4799]: I1129 04:58:12.384378 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 29 04:58:13 crc kubenswrapper[4799]: I1129 04:58:13.384252 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="4e2d8276-3c6c-418a-8d82-7b574dc1dd63" containerName="nova-metadata-log" containerID="cri-o://ae0ad8cc7e6278c605e29f443008d725401ca68b95af91b61cd7a19cbef8fa66" gracePeriod=30 Nov 29 04:58:13 crc kubenswrapper[4799]: I1129 04:58:13.384300 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="4e2d8276-3c6c-418a-8d82-7b574dc1dd63" containerName="nova-metadata-metadata" containerID="cri-o://38b6e6f85723c79313b1275c8ce3a65d703aa753ec7fb78e4971491a0cce0102" gracePeriod=30 Nov 29 04:58:14 crc kubenswrapper[4799]: I1129 04:58:14.399375 4799 generic.go:334] "Generic (PLEG): container finished" podID="4e2d8276-3c6c-418a-8d82-7b574dc1dd63" containerID="ae0ad8cc7e6278c605e29f443008d725401ca68b95af91b61cd7a19cbef8fa66" exitCode=143 Nov 29 04:58:14 crc kubenswrapper[4799]: I1129 04:58:14.399729 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4e2d8276-3c6c-418a-8d82-7b574dc1dd63","Type":"ContainerDied","Data":"ae0ad8cc7e6278c605e29f443008d725401ca68b95af91b61cd7a19cbef8fa66"} Nov 29 04:58:15 crc kubenswrapper[4799]: I1129 04:58:15.167449 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 29 04:58:15 crc kubenswrapper[4799]: I1129 04:58:15.327238 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqt5f\" (UniqueName: \"kubernetes.io/projected/26f4f01b-2064-4b8c-896e-9cf708c47ebf-kube-api-access-dqt5f\") pod \"26f4f01b-2064-4b8c-896e-9cf708c47ebf\" (UID: \"26f4f01b-2064-4b8c-896e-9cf708c47ebf\") " Nov 29 04:58:15 crc kubenswrapper[4799]: I1129 04:58:15.327367 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26f4f01b-2064-4b8c-896e-9cf708c47ebf-config-data\") pod \"26f4f01b-2064-4b8c-896e-9cf708c47ebf\" (UID: \"26f4f01b-2064-4b8c-896e-9cf708c47ebf\") " Nov 29 04:58:15 crc kubenswrapper[4799]: I1129 04:58:15.327446 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26f4f01b-2064-4b8c-896e-9cf708c47ebf-combined-ca-bundle\") pod \"26f4f01b-2064-4b8c-896e-9cf708c47ebf\" (UID: \"26f4f01b-2064-4b8c-896e-9cf708c47ebf\") " Nov 29 04:58:15 crc kubenswrapper[4799]: I1129 04:58:15.327528 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/26f4f01b-2064-4b8c-896e-9cf708c47ebf-public-tls-certs\") pod \"26f4f01b-2064-4b8c-896e-9cf708c47ebf\" (UID: \"26f4f01b-2064-4b8c-896e-9cf708c47ebf\") " Nov 29 04:58:15 crc kubenswrapper[4799]: I1129 04:58:15.328430 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/26f4f01b-2064-4b8c-896e-9cf708c47ebf-internal-tls-certs\") pod \"26f4f01b-2064-4b8c-896e-9cf708c47ebf\" (UID: \"26f4f01b-2064-4b8c-896e-9cf708c47ebf\") " Nov 29 04:58:15 crc kubenswrapper[4799]: I1129 04:58:15.328608 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26f4f01b-2064-4b8c-896e-9cf708c47ebf-logs\") pod \"26f4f01b-2064-4b8c-896e-9cf708c47ebf\" (UID: \"26f4f01b-2064-4b8c-896e-9cf708c47ebf\") " Nov 29 04:58:15 crc kubenswrapper[4799]: I1129 04:58:15.329338 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26f4f01b-2064-4b8c-896e-9cf708c47ebf-logs" (OuterVolumeSpecName: "logs") pod "26f4f01b-2064-4b8c-896e-9cf708c47ebf" (UID: "26f4f01b-2064-4b8c-896e-9cf708c47ebf"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:58:15 crc kubenswrapper[4799]: I1129 04:58:15.333331 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26f4f01b-2064-4b8c-896e-9cf708c47ebf-kube-api-access-dqt5f" (OuterVolumeSpecName: "kube-api-access-dqt5f") pod "26f4f01b-2064-4b8c-896e-9cf708c47ebf" (UID: "26f4f01b-2064-4b8c-896e-9cf708c47ebf"). InnerVolumeSpecName "kube-api-access-dqt5f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:58:15 crc kubenswrapper[4799]: I1129 04:58:15.386762 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26f4f01b-2064-4b8c-896e-9cf708c47ebf-config-data" (OuterVolumeSpecName: "config-data") pod "26f4f01b-2064-4b8c-896e-9cf708c47ebf" (UID: "26f4f01b-2064-4b8c-896e-9cf708c47ebf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:58:15 crc kubenswrapper[4799]: I1129 04:58:15.387251 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26f4f01b-2064-4b8c-896e-9cf708c47ebf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "26f4f01b-2064-4b8c-896e-9cf708c47ebf" (UID: "26f4f01b-2064-4b8c-896e-9cf708c47ebf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:58:15 crc kubenswrapper[4799]: I1129 04:58:15.412006 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26f4f01b-2064-4b8c-896e-9cf708c47ebf-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "26f4f01b-2064-4b8c-896e-9cf708c47ebf" (UID: "26f4f01b-2064-4b8c-896e-9cf708c47ebf"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:58:15 crc kubenswrapper[4799]: I1129 04:58:15.416253 4799 generic.go:334] "Generic (PLEG): container finished" podID="26f4f01b-2064-4b8c-896e-9cf708c47ebf" containerID="1e85574ad1103ec2a404861fcfc60e48c1adda765ded997e03bd6f0da1dbd49e" exitCode=0 Nov 29 04:58:15 crc kubenswrapper[4799]: I1129 04:58:15.416310 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"26f4f01b-2064-4b8c-896e-9cf708c47ebf","Type":"ContainerDied","Data":"1e85574ad1103ec2a404861fcfc60e48c1adda765ded997e03bd6f0da1dbd49e"} Nov 29 04:58:15 crc kubenswrapper[4799]: I1129 04:58:15.416342 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"26f4f01b-2064-4b8c-896e-9cf708c47ebf","Type":"ContainerDied","Data":"7e144b288b33cf1c9a2bbcbae169de58b99486ee172de18e655c1771cda8a649"} Nov 29 04:58:15 crc kubenswrapper[4799]: I1129 04:58:15.416364 4799 scope.go:117] "RemoveContainer" containerID="1e85574ad1103ec2a404861fcfc60e48c1adda765ded997e03bd6f0da1dbd49e" Nov 29 04:58:15 crc kubenswrapper[4799]: I1129 04:58:15.416652 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 29 04:58:15 crc kubenswrapper[4799]: I1129 04:58:15.427387 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26f4f01b-2064-4b8c-896e-9cf708c47ebf-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "26f4f01b-2064-4b8c-896e-9cf708c47ebf" (UID: "26f4f01b-2064-4b8c-896e-9cf708c47ebf"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:58:15 crc kubenswrapper[4799]: I1129 04:58:15.430522 4799 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/26f4f01b-2064-4b8c-896e-9cf708c47ebf-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 29 04:58:15 crc kubenswrapper[4799]: I1129 04:58:15.430562 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26f4f01b-2064-4b8c-896e-9cf708c47ebf-logs\") on node \"crc\" DevicePath \"\"" Nov 29 04:58:15 crc kubenswrapper[4799]: I1129 04:58:15.430575 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqt5f\" (UniqueName: \"kubernetes.io/projected/26f4f01b-2064-4b8c-896e-9cf708c47ebf-kube-api-access-dqt5f\") on node \"crc\" DevicePath \"\"" Nov 29 04:58:15 crc kubenswrapper[4799]: I1129 04:58:15.430590 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26f4f01b-2064-4b8c-896e-9cf708c47ebf-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 04:58:15 crc kubenswrapper[4799]: I1129 04:58:15.430602 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26f4f01b-2064-4b8c-896e-9cf708c47ebf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 04:58:15 crc kubenswrapper[4799]: I1129 04:58:15.430613 4799 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/26f4f01b-2064-4b8c-896e-9cf708c47ebf-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 29 04:58:15 crc kubenswrapper[4799]: I1129 04:58:15.447592 4799 scope.go:117] "RemoveContainer" containerID="dd097e8aed8ba6307b766d41135154ddeee364b896e797b1d53adbd27bae566d" Nov 29 04:58:15 crc kubenswrapper[4799]: I1129 04:58:15.467234 4799 scope.go:117] "RemoveContainer" containerID="1e85574ad1103ec2a404861fcfc60e48c1adda765ded997e03bd6f0da1dbd49e" Nov 29 04:58:15 crc kubenswrapper[4799]: E1129 04:58:15.467625 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e85574ad1103ec2a404861fcfc60e48c1adda765ded997e03bd6f0da1dbd49e\": container with ID starting with 1e85574ad1103ec2a404861fcfc60e48c1adda765ded997e03bd6f0da1dbd49e not found: ID does not exist" containerID="1e85574ad1103ec2a404861fcfc60e48c1adda765ded997e03bd6f0da1dbd49e" Nov 29 04:58:15 crc kubenswrapper[4799]: I1129 04:58:15.467659 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e85574ad1103ec2a404861fcfc60e48c1adda765ded997e03bd6f0da1dbd49e"} err="failed to get container status \"1e85574ad1103ec2a404861fcfc60e48c1adda765ded997e03bd6f0da1dbd49e\": rpc error: code = NotFound desc = could not find container \"1e85574ad1103ec2a404861fcfc60e48c1adda765ded997e03bd6f0da1dbd49e\": container with ID starting with 1e85574ad1103ec2a404861fcfc60e48c1adda765ded997e03bd6f0da1dbd49e not found: ID does not exist" Nov 29 04:58:15 crc kubenswrapper[4799]: I1129 04:58:15.467680 4799 scope.go:117] "RemoveContainer" containerID="dd097e8aed8ba6307b766d41135154ddeee364b896e797b1d53adbd27bae566d" Nov 29 04:58:15 crc kubenswrapper[4799]: E1129 04:58:15.467962 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd097e8aed8ba6307b766d41135154ddeee364b896e797b1d53adbd27bae566d\": container with ID starting with dd097e8aed8ba6307b766d41135154ddeee364b896e797b1d53adbd27bae566d not found: ID does not exist" containerID="dd097e8aed8ba6307b766d41135154ddeee364b896e797b1d53adbd27bae566d" Nov 29 04:58:15 crc kubenswrapper[4799]: I1129 04:58:15.468007 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd097e8aed8ba6307b766d41135154ddeee364b896e797b1d53adbd27bae566d"} err="failed to get container status \"dd097e8aed8ba6307b766d41135154ddeee364b896e797b1d53adbd27bae566d\": rpc error: code = NotFound desc = could not find container \"dd097e8aed8ba6307b766d41135154ddeee364b896e797b1d53adbd27bae566d\": container with ID starting with dd097e8aed8ba6307b766d41135154ddeee364b896e797b1d53adbd27bae566d not found: ID does not exist" Nov 29 04:58:15 crc kubenswrapper[4799]: I1129 04:58:15.751933 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 29 04:58:15 crc kubenswrapper[4799]: I1129 04:58:15.761387 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 29 04:58:15 crc kubenswrapper[4799]: I1129 04:58:15.783005 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 29 04:58:15 crc kubenswrapper[4799]: E1129 04:58:15.783525 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26f4f01b-2064-4b8c-896e-9cf708c47ebf" containerName="nova-api-log" Nov 29 04:58:15 crc kubenswrapper[4799]: I1129 04:58:15.783552 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="26f4f01b-2064-4b8c-896e-9cf708c47ebf" containerName="nova-api-log" Nov 29 04:58:15 crc kubenswrapper[4799]: E1129 04:58:15.783572 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35dd3ae0-d97c-49ff-8b70-a394fd58af0f" containerName="nova-manage" Nov 29 04:58:15 crc kubenswrapper[4799]: I1129 04:58:15.783581 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="35dd3ae0-d97c-49ff-8b70-a394fd58af0f" containerName="nova-manage" Nov 29 04:58:15 crc kubenswrapper[4799]: E1129 04:58:15.783623 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26f4f01b-2064-4b8c-896e-9cf708c47ebf" containerName="nova-api-api" Nov 29 04:58:15 crc kubenswrapper[4799]: I1129 04:58:15.783635 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="26f4f01b-2064-4b8c-896e-9cf708c47ebf" containerName="nova-api-api" Nov 29 04:58:15 crc kubenswrapper[4799]: I1129 04:58:15.783880 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="26f4f01b-2064-4b8c-896e-9cf708c47ebf" containerName="nova-api-api" Nov 29 04:58:15 crc kubenswrapper[4799]: I1129 04:58:15.783909 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="35dd3ae0-d97c-49ff-8b70-a394fd58af0f" containerName="nova-manage" Nov 29 04:58:15 crc kubenswrapper[4799]: I1129 04:58:15.783940 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="26f4f01b-2064-4b8c-896e-9cf708c47ebf" containerName="nova-api-log" Nov 29 04:58:15 crc kubenswrapper[4799]: I1129 04:58:15.785254 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 29 04:58:15 crc kubenswrapper[4799]: I1129 04:58:15.789408 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 29 04:58:15 crc kubenswrapper[4799]: I1129 04:58:15.804245 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 29 04:58:15 crc kubenswrapper[4799]: I1129 04:58:15.805181 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 29 04:58:15 crc kubenswrapper[4799]: I1129 04:58:15.819086 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 29 04:58:15 crc kubenswrapper[4799]: I1129 04:58:15.940918 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd1dc76e-76ba-4ab4-b40c-587ffbf9b5df-config-data\") pod \"nova-api-0\" (UID: \"cd1dc76e-76ba-4ab4-b40c-587ffbf9b5df\") " pod="openstack/nova-api-0" Nov 29 04:58:15 crc kubenswrapper[4799]: I1129 04:58:15.941266 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd1dc76e-76ba-4ab4-b40c-587ffbf9b5df-public-tls-certs\") pod \"nova-api-0\" (UID: \"cd1dc76e-76ba-4ab4-b40c-587ffbf9b5df\") " pod="openstack/nova-api-0" Nov 29 04:58:15 crc kubenswrapper[4799]: I1129 04:58:15.941440 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd1dc76e-76ba-4ab4-b40c-587ffbf9b5df-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cd1dc76e-76ba-4ab4-b40c-587ffbf9b5df\") " pod="openstack/nova-api-0" Nov 29 04:58:15 crc kubenswrapper[4799]: I1129 04:58:15.941490 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zddg\" (UniqueName: \"kubernetes.io/projected/cd1dc76e-76ba-4ab4-b40c-587ffbf9b5df-kube-api-access-5zddg\") pod \"nova-api-0\" (UID: \"cd1dc76e-76ba-4ab4-b40c-587ffbf9b5df\") " pod="openstack/nova-api-0" Nov 29 04:58:15 crc kubenswrapper[4799]: I1129 04:58:15.941518 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd1dc76e-76ba-4ab4-b40c-587ffbf9b5df-logs\") pod \"nova-api-0\" (UID: \"cd1dc76e-76ba-4ab4-b40c-587ffbf9b5df\") " pod="openstack/nova-api-0" Nov 29 04:58:15 crc kubenswrapper[4799]: I1129 04:58:15.941542 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd1dc76e-76ba-4ab4-b40c-587ffbf9b5df-internal-tls-certs\") pod \"nova-api-0\" (UID: \"cd1dc76e-76ba-4ab4-b40c-587ffbf9b5df\") " pod="openstack/nova-api-0" Nov 29 04:58:16 crc kubenswrapper[4799]: I1129 04:58:16.043796 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd1dc76e-76ba-4ab4-b40c-587ffbf9b5df-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cd1dc76e-76ba-4ab4-b40c-587ffbf9b5df\") " pod="openstack/nova-api-0" Nov 29 04:58:16 crc kubenswrapper[4799]: I1129 04:58:16.043852 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zddg\" (UniqueName: \"kubernetes.io/projected/cd1dc76e-76ba-4ab4-b40c-587ffbf9b5df-kube-api-access-5zddg\") pod \"nova-api-0\" (UID: \"cd1dc76e-76ba-4ab4-b40c-587ffbf9b5df\") " pod="openstack/nova-api-0" Nov 29 04:58:16 crc kubenswrapper[4799]: I1129 04:58:16.043871 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd1dc76e-76ba-4ab4-b40c-587ffbf9b5df-internal-tls-certs\") pod \"nova-api-0\" (UID: \"cd1dc76e-76ba-4ab4-b40c-587ffbf9b5df\") " pod="openstack/nova-api-0" Nov 29 04:58:16 crc kubenswrapper[4799]: I1129 04:58:16.043892 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd1dc76e-76ba-4ab4-b40c-587ffbf9b5df-logs\") pod \"nova-api-0\" (UID: \"cd1dc76e-76ba-4ab4-b40c-587ffbf9b5df\") " pod="openstack/nova-api-0" Nov 29 04:58:16 crc kubenswrapper[4799]: I1129 04:58:16.043968 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd1dc76e-76ba-4ab4-b40c-587ffbf9b5df-config-data\") pod \"nova-api-0\" (UID: \"cd1dc76e-76ba-4ab4-b40c-587ffbf9b5df\") " pod="openstack/nova-api-0" Nov 29 04:58:16 crc kubenswrapper[4799]: I1129 04:58:16.043990 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd1dc76e-76ba-4ab4-b40c-587ffbf9b5df-public-tls-certs\") pod \"nova-api-0\" (UID: \"cd1dc76e-76ba-4ab4-b40c-587ffbf9b5df\") " pod="openstack/nova-api-0" Nov 29 04:58:16 crc kubenswrapper[4799]: I1129 04:58:16.044879 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd1dc76e-76ba-4ab4-b40c-587ffbf9b5df-logs\") pod \"nova-api-0\" (UID: \"cd1dc76e-76ba-4ab4-b40c-587ffbf9b5df\") " pod="openstack/nova-api-0" Nov 29 04:58:16 crc kubenswrapper[4799]: I1129 04:58:16.051087 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd1dc76e-76ba-4ab4-b40c-587ffbf9b5df-internal-tls-certs\") pod \"nova-api-0\" (UID: \"cd1dc76e-76ba-4ab4-b40c-587ffbf9b5df\") " pod="openstack/nova-api-0" Nov 29 04:58:16 crc kubenswrapper[4799]: I1129 04:58:16.051179 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd1dc76e-76ba-4ab4-b40c-587ffbf9b5df-config-data\") pod \"nova-api-0\" (UID: \"cd1dc76e-76ba-4ab4-b40c-587ffbf9b5df\") " pod="openstack/nova-api-0" Nov 29 04:58:16 crc kubenswrapper[4799]: I1129 04:58:16.053221 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd1dc76e-76ba-4ab4-b40c-587ffbf9b5df-public-tls-certs\") pod \"nova-api-0\" (UID: \"cd1dc76e-76ba-4ab4-b40c-587ffbf9b5df\") " pod="openstack/nova-api-0" Nov 29 04:58:16 crc kubenswrapper[4799]: I1129 04:58:16.053474 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd1dc76e-76ba-4ab4-b40c-587ffbf9b5df-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cd1dc76e-76ba-4ab4-b40c-587ffbf9b5df\") " pod="openstack/nova-api-0" Nov 29 04:58:16 crc kubenswrapper[4799]: I1129 04:58:16.063504 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zddg\" (UniqueName: \"kubernetes.io/projected/cd1dc76e-76ba-4ab4-b40c-587ffbf9b5df-kube-api-access-5zddg\") pod \"nova-api-0\" (UID: \"cd1dc76e-76ba-4ab4-b40c-587ffbf9b5df\") " pod="openstack/nova-api-0" Nov 29 04:58:16 crc kubenswrapper[4799]: I1129 04:58:16.170323 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 29 04:58:16 crc kubenswrapper[4799]: I1129 04:58:16.367206 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 29 04:58:16 crc kubenswrapper[4799]: I1129 04:58:16.437203 4799 generic.go:334] "Generic (PLEG): container finished" podID="3c25c970-ef30-438b-b5d3-a98976fa7caa" containerID="1879026d9dc12823fb9939d498c22c08936c600c692b8f34e8b5385e53f855f4" exitCode=0 Nov 29 04:58:16 crc kubenswrapper[4799]: I1129 04:58:16.437317 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 29 04:58:16 crc kubenswrapper[4799]: I1129 04:58:16.437295 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3c25c970-ef30-438b-b5d3-a98976fa7caa","Type":"ContainerDied","Data":"1879026d9dc12823fb9939d498c22c08936c600c692b8f34e8b5385e53f855f4"} Nov 29 04:58:16 crc kubenswrapper[4799]: I1129 04:58:16.437486 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3c25c970-ef30-438b-b5d3-a98976fa7caa","Type":"ContainerDied","Data":"551552a170f802f9a09bcba020727ba2583db85271234e8c871ca4157602c855"} Nov 29 04:58:16 crc kubenswrapper[4799]: I1129 04:58:16.437536 4799 scope.go:117] "RemoveContainer" containerID="1879026d9dc12823fb9939d498c22c08936c600c692b8f34e8b5385e53f855f4" Nov 29 04:58:16 crc kubenswrapper[4799]: I1129 04:58:16.452913 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zhknj\" (UniqueName: \"kubernetes.io/projected/3c25c970-ef30-438b-b5d3-a98976fa7caa-kube-api-access-zhknj\") pod \"3c25c970-ef30-438b-b5d3-a98976fa7caa\" (UID: \"3c25c970-ef30-438b-b5d3-a98976fa7caa\") " Nov 29 04:58:16 crc kubenswrapper[4799]: I1129 04:58:16.453117 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c25c970-ef30-438b-b5d3-a98976fa7caa-config-data\") pod \"3c25c970-ef30-438b-b5d3-a98976fa7caa\" (UID: \"3c25c970-ef30-438b-b5d3-a98976fa7caa\") " Nov 29 04:58:16 crc kubenswrapper[4799]: I1129 04:58:16.453305 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c25c970-ef30-438b-b5d3-a98976fa7caa-combined-ca-bundle\") pod \"3c25c970-ef30-438b-b5d3-a98976fa7caa\" (UID: \"3c25c970-ef30-438b-b5d3-a98976fa7caa\") " Nov 29 04:58:16 crc kubenswrapper[4799]: I1129 04:58:16.457942 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c25c970-ef30-438b-b5d3-a98976fa7caa-kube-api-access-zhknj" (OuterVolumeSpecName: "kube-api-access-zhknj") pod "3c25c970-ef30-438b-b5d3-a98976fa7caa" (UID: "3c25c970-ef30-438b-b5d3-a98976fa7caa"). InnerVolumeSpecName "kube-api-access-zhknj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:58:16 crc kubenswrapper[4799]: I1129 04:58:16.475074 4799 scope.go:117] "RemoveContainer" containerID="1879026d9dc12823fb9939d498c22c08936c600c692b8f34e8b5385e53f855f4" Nov 29 04:58:16 crc kubenswrapper[4799]: E1129 04:58:16.477093 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1879026d9dc12823fb9939d498c22c08936c600c692b8f34e8b5385e53f855f4\": container with ID starting with 1879026d9dc12823fb9939d498c22c08936c600c692b8f34e8b5385e53f855f4 not found: ID does not exist" containerID="1879026d9dc12823fb9939d498c22c08936c600c692b8f34e8b5385e53f855f4" Nov 29 04:58:16 crc kubenswrapper[4799]: I1129 04:58:16.477156 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1879026d9dc12823fb9939d498c22c08936c600c692b8f34e8b5385e53f855f4"} err="failed to get container status \"1879026d9dc12823fb9939d498c22c08936c600c692b8f34e8b5385e53f855f4\": rpc error: code = NotFound desc = could not find container \"1879026d9dc12823fb9939d498c22c08936c600c692b8f34e8b5385e53f855f4\": container with ID starting with 1879026d9dc12823fb9939d498c22c08936c600c692b8f34e8b5385e53f855f4 not found: ID does not exist" Nov 29 04:58:16 crc kubenswrapper[4799]: I1129 04:58:16.484831 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c25c970-ef30-438b-b5d3-a98976fa7caa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3c25c970-ef30-438b-b5d3-a98976fa7caa" (UID: "3c25c970-ef30-438b-b5d3-a98976fa7caa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:58:16 crc kubenswrapper[4799]: I1129 04:58:16.485223 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c25c970-ef30-438b-b5d3-a98976fa7caa-config-data" (OuterVolumeSpecName: "config-data") pod "3c25c970-ef30-438b-b5d3-a98976fa7caa" (UID: "3c25c970-ef30-438b-b5d3-a98976fa7caa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:58:16 crc kubenswrapper[4799]: I1129 04:58:16.530566 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="4e2d8276-3c6c-418a-8d82-7b574dc1dd63" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.178:8775/\": dial tcp 10.217.0.178:8775: connect: connection refused" Nov 29 04:58:16 crc kubenswrapper[4799]: I1129 04:58:16.530946 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="4e2d8276-3c6c-418a-8d82-7b574dc1dd63" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.178:8775/\": dial tcp 10.217.0.178:8775: connect: connection refused" Nov 29 04:58:16 crc kubenswrapper[4799]: I1129 04:58:16.555368 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c25c970-ef30-438b-b5d3-a98976fa7caa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 04:58:16 crc kubenswrapper[4799]: I1129 04:58:16.555410 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zhknj\" (UniqueName: \"kubernetes.io/projected/3c25c970-ef30-438b-b5d3-a98976fa7caa-kube-api-access-zhknj\") on node \"crc\" DevicePath \"\"" Nov 29 04:58:16 crc kubenswrapper[4799]: I1129 04:58:16.555426 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c25c970-ef30-438b-b5d3-a98976fa7caa-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 04:58:16 crc kubenswrapper[4799]: I1129 04:58:16.678620 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26f4f01b-2064-4b8c-896e-9cf708c47ebf" path="/var/lib/kubelet/pods/26f4f01b-2064-4b8c-896e-9cf708c47ebf/volumes" Nov 29 04:58:16 crc kubenswrapper[4799]: I1129 04:58:16.689469 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 29 04:58:16 crc kubenswrapper[4799]: W1129 04:58:16.689894 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd1dc76e_76ba_4ab4_b40c_587ffbf9b5df.slice/crio-6fc955ed67c324a14ae4fd860be72534681db0b76221e62d18a50e1d918e5903 WatchSource:0}: Error finding container 6fc955ed67c324a14ae4fd860be72534681db0b76221e62d18a50e1d918e5903: Status 404 returned error can't find the container with id 6fc955ed67c324a14ae4fd860be72534681db0b76221e62d18a50e1d918e5903 Nov 29 04:58:16 crc kubenswrapper[4799]: I1129 04:58:16.797579 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 29 04:58:16 crc kubenswrapper[4799]: I1129 04:58:16.830469 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 29 04:58:16 crc kubenswrapper[4799]: I1129 04:58:16.852563 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 29 04:58:16 crc kubenswrapper[4799]: E1129 04:58:16.853488 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c25c970-ef30-438b-b5d3-a98976fa7caa" containerName="nova-scheduler-scheduler" Nov 29 04:58:16 crc kubenswrapper[4799]: I1129 04:58:16.853516 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c25c970-ef30-438b-b5d3-a98976fa7caa" containerName="nova-scheduler-scheduler" Nov 29 04:58:16 crc kubenswrapper[4799]: I1129 04:58:16.853749 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c25c970-ef30-438b-b5d3-a98976fa7caa" containerName="nova-scheduler-scheduler" Nov 29 04:58:16 crc kubenswrapper[4799]: I1129 04:58:16.854700 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 29 04:58:16 crc kubenswrapper[4799]: I1129 04:58:16.857776 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 29 04:58:16 crc kubenswrapper[4799]: I1129 04:58:16.872036 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 29 04:58:16 crc kubenswrapper[4799]: I1129 04:58:16.964406 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwls5\" (UniqueName: \"kubernetes.io/projected/923b93cd-9a7f-4924-8d46-a1cd33612af5-kube-api-access-kwls5\") pod \"nova-scheduler-0\" (UID: \"923b93cd-9a7f-4924-8d46-a1cd33612af5\") " pod="openstack/nova-scheduler-0" Nov 29 04:58:16 crc kubenswrapper[4799]: I1129 04:58:16.964525 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/923b93cd-9a7f-4924-8d46-a1cd33612af5-config-data\") pod \"nova-scheduler-0\" (UID: \"923b93cd-9a7f-4924-8d46-a1cd33612af5\") " pod="openstack/nova-scheduler-0" Nov 29 04:58:16 crc kubenswrapper[4799]: I1129 04:58:16.964605 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/923b93cd-9a7f-4924-8d46-a1cd33612af5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"923b93cd-9a7f-4924-8d46-a1cd33612af5\") " pod="openstack/nova-scheduler-0" Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.046278 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.076643 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwls5\" (UniqueName: \"kubernetes.io/projected/923b93cd-9a7f-4924-8d46-a1cd33612af5-kube-api-access-kwls5\") pod \"nova-scheduler-0\" (UID: \"923b93cd-9a7f-4924-8d46-a1cd33612af5\") " pod="openstack/nova-scheduler-0" Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.076744 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/923b93cd-9a7f-4924-8d46-a1cd33612af5-config-data\") pod \"nova-scheduler-0\" (UID: \"923b93cd-9a7f-4924-8d46-a1cd33612af5\") " pod="openstack/nova-scheduler-0" Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.076977 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/923b93cd-9a7f-4924-8d46-a1cd33612af5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"923b93cd-9a7f-4924-8d46-a1cd33612af5\") " pod="openstack/nova-scheduler-0" Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.090004 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/923b93cd-9a7f-4924-8d46-a1cd33612af5-config-data\") pod \"nova-scheduler-0\" (UID: \"923b93cd-9a7f-4924-8d46-a1cd33612af5\") " pod="openstack/nova-scheduler-0" Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.100734 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/923b93cd-9a7f-4924-8d46-a1cd33612af5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"923b93cd-9a7f-4924-8d46-a1cd33612af5\") " pod="openstack/nova-scheduler-0" Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.124044 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwls5\" (UniqueName: \"kubernetes.io/projected/923b93cd-9a7f-4924-8d46-a1cd33612af5-kube-api-access-kwls5\") pod \"nova-scheduler-0\" (UID: \"923b93cd-9a7f-4924-8d46-a1cd33612af5\") " pod="openstack/nova-scheduler-0" Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.178774 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xb8rj\" (UniqueName: \"kubernetes.io/projected/4e2d8276-3c6c-418a-8d82-7b574dc1dd63-kube-api-access-xb8rj\") pod \"4e2d8276-3c6c-418a-8d82-7b574dc1dd63\" (UID: \"4e2d8276-3c6c-418a-8d82-7b574dc1dd63\") " Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.178865 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e2d8276-3c6c-418a-8d82-7b574dc1dd63-nova-metadata-tls-certs\") pod \"4e2d8276-3c6c-418a-8d82-7b574dc1dd63\" (UID: \"4e2d8276-3c6c-418a-8d82-7b574dc1dd63\") " Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.179030 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e2d8276-3c6c-418a-8d82-7b574dc1dd63-combined-ca-bundle\") pod \"4e2d8276-3c6c-418a-8d82-7b574dc1dd63\" (UID: \"4e2d8276-3c6c-418a-8d82-7b574dc1dd63\") " Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.179076 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e2d8276-3c6c-418a-8d82-7b574dc1dd63-config-data\") pod \"4e2d8276-3c6c-418a-8d82-7b574dc1dd63\" (UID: \"4e2d8276-3c6c-418a-8d82-7b574dc1dd63\") " Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.179102 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e2d8276-3c6c-418a-8d82-7b574dc1dd63-logs\") pod \"4e2d8276-3c6c-418a-8d82-7b574dc1dd63\" (UID: \"4e2d8276-3c6c-418a-8d82-7b574dc1dd63\") " Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.179770 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e2d8276-3c6c-418a-8d82-7b574dc1dd63-logs" (OuterVolumeSpecName: "logs") pod "4e2d8276-3c6c-418a-8d82-7b574dc1dd63" (UID: "4e2d8276-3c6c-418a-8d82-7b574dc1dd63"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.182715 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.183356 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e2d8276-3c6c-418a-8d82-7b574dc1dd63-kube-api-access-xb8rj" (OuterVolumeSpecName: "kube-api-access-xb8rj") pod "4e2d8276-3c6c-418a-8d82-7b574dc1dd63" (UID: "4e2d8276-3c6c-418a-8d82-7b574dc1dd63"). InnerVolumeSpecName "kube-api-access-xb8rj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.209818 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e2d8276-3c6c-418a-8d82-7b574dc1dd63-config-data" (OuterVolumeSpecName: "config-data") pod "4e2d8276-3c6c-418a-8d82-7b574dc1dd63" (UID: "4e2d8276-3c6c-418a-8d82-7b574dc1dd63"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.216295 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e2d8276-3c6c-418a-8d82-7b574dc1dd63-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4e2d8276-3c6c-418a-8d82-7b574dc1dd63" (UID: "4e2d8276-3c6c-418a-8d82-7b574dc1dd63"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.249022 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e2d8276-3c6c-418a-8d82-7b574dc1dd63-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "4e2d8276-3c6c-418a-8d82-7b574dc1dd63" (UID: "4e2d8276-3c6c-418a-8d82-7b574dc1dd63"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.288598 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xb8rj\" (UniqueName: \"kubernetes.io/projected/4e2d8276-3c6c-418a-8d82-7b574dc1dd63-kube-api-access-xb8rj\") on node \"crc\" DevicePath \"\"" Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.288987 4799 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e2d8276-3c6c-418a-8d82-7b574dc1dd63-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.289002 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e2d8276-3c6c-418a-8d82-7b574dc1dd63-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.289011 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e2d8276-3c6c-418a-8d82-7b574dc1dd63-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.289021 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e2d8276-3c6c-418a-8d82-7b574dc1dd63-logs\") on node \"crc\" DevicePath \"\"" Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.457166 4799 generic.go:334] "Generic (PLEG): container finished" podID="4e2d8276-3c6c-418a-8d82-7b574dc1dd63" containerID="38b6e6f85723c79313b1275c8ce3a65d703aa753ec7fb78e4971491a0cce0102" exitCode=0 Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.457302 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.457317 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4e2d8276-3c6c-418a-8d82-7b574dc1dd63","Type":"ContainerDied","Data":"38b6e6f85723c79313b1275c8ce3a65d703aa753ec7fb78e4971491a0cce0102"} Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.457535 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4e2d8276-3c6c-418a-8d82-7b574dc1dd63","Type":"ContainerDied","Data":"2a43d4c7c1590b1a1006eb0d7782448217401cb2fee1ec735600f117684db004"} Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.457563 4799 scope.go:117] "RemoveContainer" containerID="38b6e6f85723c79313b1275c8ce3a65d703aa753ec7fb78e4971491a0cce0102" Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.471297 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cd1dc76e-76ba-4ab4-b40c-587ffbf9b5df","Type":"ContainerStarted","Data":"b101593f0db0ec4bf64fa2665cfd016781f5fc950765d8515379b682704b0c2f"} Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.471370 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cd1dc76e-76ba-4ab4-b40c-587ffbf9b5df","Type":"ContainerStarted","Data":"6fc955ed67c324a14ae4fd860be72534681db0b76221e62d18a50e1d918e5903"} Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.490521 4799 scope.go:117] "RemoveContainer" containerID="ae0ad8cc7e6278c605e29f443008d725401ca68b95af91b61cd7a19cbef8fa66" Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.502973 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.516649 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.526418 4799 scope.go:117] "RemoveContainer" containerID="38b6e6f85723c79313b1275c8ce3a65d703aa753ec7fb78e4971491a0cce0102" Nov 29 04:58:17 crc kubenswrapper[4799]: E1129 04:58:17.527469 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38b6e6f85723c79313b1275c8ce3a65d703aa753ec7fb78e4971491a0cce0102\": container with ID starting with 38b6e6f85723c79313b1275c8ce3a65d703aa753ec7fb78e4971491a0cce0102 not found: ID does not exist" containerID="38b6e6f85723c79313b1275c8ce3a65d703aa753ec7fb78e4971491a0cce0102" Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.527518 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38b6e6f85723c79313b1275c8ce3a65d703aa753ec7fb78e4971491a0cce0102"} err="failed to get container status \"38b6e6f85723c79313b1275c8ce3a65d703aa753ec7fb78e4971491a0cce0102\": rpc error: code = NotFound desc = could not find container \"38b6e6f85723c79313b1275c8ce3a65d703aa753ec7fb78e4971491a0cce0102\": container with ID starting with 38b6e6f85723c79313b1275c8ce3a65d703aa753ec7fb78e4971491a0cce0102 not found: ID does not exist" Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.527575 4799 scope.go:117] "RemoveContainer" containerID="ae0ad8cc7e6278c605e29f443008d725401ca68b95af91b61cd7a19cbef8fa66" Nov 29 04:58:17 crc kubenswrapper[4799]: E1129 04:58:17.528176 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae0ad8cc7e6278c605e29f443008d725401ca68b95af91b61cd7a19cbef8fa66\": container with ID starting with ae0ad8cc7e6278c605e29f443008d725401ca68b95af91b61cd7a19cbef8fa66 not found: ID does not exist" containerID="ae0ad8cc7e6278c605e29f443008d725401ca68b95af91b61cd7a19cbef8fa66" Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.528202 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae0ad8cc7e6278c605e29f443008d725401ca68b95af91b61cd7a19cbef8fa66"} err="failed to get container status \"ae0ad8cc7e6278c605e29f443008d725401ca68b95af91b61cd7a19cbef8fa66\": rpc error: code = NotFound desc = could not find container \"ae0ad8cc7e6278c605e29f443008d725401ca68b95af91b61cd7a19cbef8fa66\": container with ID starting with ae0ad8cc7e6278c605e29f443008d725401ca68b95af91b61cd7a19cbef8fa66 not found: ID does not exist" Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.532384 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 29 04:58:17 crc kubenswrapper[4799]: E1129 04:58:17.532865 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e2d8276-3c6c-418a-8d82-7b574dc1dd63" containerName="nova-metadata-metadata" Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.532911 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e2d8276-3c6c-418a-8d82-7b574dc1dd63" containerName="nova-metadata-metadata" Nov 29 04:58:17 crc kubenswrapper[4799]: E1129 04:58:17.532965 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e2d8276-3c6c-418a-8d82-7b574dc1dd63" containerName="nova-metadata-log" Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.532972 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e2d8276-3c6c-418a-8d82-7b574dc1dd63" containerName="nova-metadata-log" Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.533214 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e2d8276-3c6c-418a-8d82-7b574dc1dd63" containerName="nova-metadata-log" Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.533241 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e2d8276-3c6c-418a-8d82-7b574dc1dd63" containerName="nova-metadata-metadata" Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.534479 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.537476 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.547436 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.553062 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.677176 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 29 04:58:17 crc kubenswrapper[4799]: W1129 04:58:17.684078 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod923b93cd_9a7f_4924_8d46_a1cd33612af5.slice/crio-011b1dbff02fe537fa8644e477759cabdce57a19bed4b54fca4385227062f1b7 WatchSource:0}: Error finding container 011b1dbff02fe537fa8644e477759cabdce57a19bed4b54fca4385227062f1b7: Status 404 returned error can't find the container with id 011b1dbff02fe537fa8644e477759cabdce57a19bed4b54fca4385227062f1b7 Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.712207 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfbbd39d-1e8f-4bdd-991e-71e206bb4ca0-config-data\") pod \"nova-metadata-0\" (UID: \"bfbbd39d-1e8f-4bdd-991e-71e206bb4ca0\") " pod="openstack/nova-metadata-0" Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.712652 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bfbbd39d-1e8f-4bdd-991e-71e206bb4ca0-logs\") pod \"nova-metadata-0\" (UID: \"bfbbd39d-1e8f-4bdd-991e-71e206bb4ca0\") " pod="openstack/nova-metadata-0" Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.712912 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfbbd39d-1e8f-4bdd-991e-71e206bb4ca0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bfbbd39d-1e8f-4bdd-991e-71e206bb4ca0\") " pod="openstack/nova-metadata-0" Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.712976 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfbbd39d-1e8f-4bdd-991e-71e206bb4ca0-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"bfbbd39d-1e8f-4bdd-991e-71e206bb4ca0\") " pod="openstack/nova-metadata-0" Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.713006 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6w2p\" (UniqueName: \"kubernetes.io/projected/bfbbd39d-1e8f-4bdd-991e-71e206bb4ca0-kube-api-access-b6w2p\") pod \"nova-metadata-0\" (UID: \"bfbbd39d-1e8f-4bdd-991e-71e206bb4ca0\") " pod="openstack/nova-metadata-0" Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.814934 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfbbd39d-1e8f-4bdd-991e-71e206bb4ca0-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"bfbbd39d-1e8f-4bdd-991e-71e206bb4ca0\") " pod="openstack/nova-metadata-0" Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.815757 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6w2p\" (UniqueName: \"kubernetes.io/projected/bfbbd39d-1e8f-4bdd-991e-71e206bb4ca0-kube-api-access-b6w2p\") pod \"nova-metadata-0\" (UID: \"bfbbd39d-1e8f-4bdd-991e-71e206bb4ca0\") " pod="openstack/nova-metadata-0" Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.815968 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfbbd39d-1e8f-4bdd-991e-71e206bb4ca0-config-data\") pod \"nova-metadata-0\" (UID: \"bfbbd39d-1e8f-4bdd-991e-71e206bb4ca0\") " pod="openstack/nova-metadata-0" Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.816059 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bfbbd39d-1e8f-4bdd-991e-71e206bb4ca0-logs\") pod \"nova-metadata-0\" (UID: \"bfbbd39d-1e8f-4bdd-991e-71e206bb4ca0\") " pod="openstack/nova-metadata-0" Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.816208 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfbbd39d-1e8f-4bdd-991e-71e206bb4ca0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bfbbd39d-1e8f-4bdd-991e-71e206bb4ca0\") " pod="openstack/nova-metadata-0" Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.816560 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bfbbd39d-1e8f-4bdd-991e-71e206bb4ca0-logs\") pod \"nova-metadata-0\" (UID: \"bfbbd39d-1e8f-4bdd-991e-71e206bb4ca0\") " pod="openstack/nova-metadata-0" Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.819919 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfbbd39d-1e8f-4bdd-991e-71e206bb4ca0-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"bfbbd39d-1e8f-4bdd-991e-71e206bb4ca0\") " pod="openstack/nova-metadata-0" Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.820633 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfbbd39d-1e8f-4bdd-991e-71e206bb4ca0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bfbbd39d-1e8f-4bdd-991e-71e206bb4ca0\") " pod="openstack/nova-metadata-0" Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.822461 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfbbd39d-1e8f-4bdd-991e-71e206bb4ca0-config-data\") pod \"nova-metadata-0\" (UID: \"bfbbd39d-1e8f-4bdd-991e-71e206bb4ca0\") " pod="openstack/nova-metadata-0" Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.842198 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6w2p\" (UniqueName: \"kubernetes.io/projected/bfbbd39d-1e8f-4bdd-991e-71e206bb4ca0-kube-api-access-b6w2p\") pod \"nova-metadata-0\" (UID: \"bfbbd39d-1e8f-4bdd-991e-71e206bb4ca0\") " pod="openstack/nova-metadata-0" Nov 29 04:58:17 crc kubenswrapper[4799]: I1129 04:58:17.855951 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 29 04:58:18 crc kubenswrapper[4799]: I1129 04:58:18.329843 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 29 04:58:18 crc kubenswrapper[4799]: I1129 04:58:18.487834 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bfbbd39d-1e8f-4bdd-991e-71e206bb4ca0","Type":"ContainerStarted","Data":"9ff430560ba807a8e0d902eb74e219834675459affcd473378bd5f9c182dfdeb"} Nov 29 04:58:18 crc kubenswrapper[4799]: I1129 04:58:18.491844 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"923b93cd-9a7f-4924-8d46-a1cd33612af5","Type":"ContainerStarted","Data":"3df9a6046b9e8d55c9a3bb97896c6b50283913ad74dc55e1af3c7aec41ac3b05"} Nov 29 04:58:18 crc kubenswrapper[4799]: I1129 04:58:18.491896 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"923b93cd-9a7f-4924-8d46-a1cd33612af5","Type":"ContainerStarted","Data":"011b1dbff02fe537fa8644e477759cabdce57a19bed4b54fca4385227062f1b7"} Nov 29 04:58:18 crc kubenswrapper[4799]: I1129 04:58:18.493876 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cd1dc76e-76ba-4ab4-b40c-587ffbf9b5df","Type":"ContainerStarted","Data":"d9899e12ea75ddff2309467d5cd10d94ba251c448e298b995ed9cdc4168a4ba0"} Nov 29 04:58:18 crc kubenswrapper[4799]: I1129 04:58:18.520681 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.520656152 podStartE2EDuration="2.520656152s" podCreationTimestamp="2025-11-29 04:58:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:58:18.512272312 +0000 UTC m=+1174.155202712" watchObservedRunningTime="2025-11-29 04:58:18.520656152 +0000 UTC m=+1174.163586552" Nov 29 04:58:18 crc kubenswrapper[4799]: I1129 04:58:18.542503 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.542480922 podStartE2EDuration="3.542480922s" podCreationTimestamp="2025-11-29 04:58:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:58:18.533298323 +0000 UTC m=+1174.176228733" watchObservedRunningTime="2025-11-29 04:58:18.542480922 +0000 UTC m=+1174.185411322" Nov 29 04:58:18 crc kubenswrapper[4799]: I1129 04:58:18.684726 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c25c970-ef30-438b-b5d3-a98976fa7caa" path="/var/lib/kubelet/pods/3c25c970-ef30-438b-b5d3-a98976fa7caa/volumes" Nov 29 04:58:18 crc kubenswrapper[4799]: I1129 04:58:18.697379 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e2d8276-3c6c-418a-8d82-7b574dc1dd63" path="/var/lib/kubelet/pods/4e2d8276-3c6c-418a-8d82-7b574dc1dd63/volumes" Nov 29 04:58:19 crc kubenswrapper[4799]: I1129 04:58:19.504812 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bfbbd39d-1e8f-4bdd-991e-71e206bb4ca0","Type":"ContainerStarted","Data":"46e67bd1b1c15bb0eacd54eb2bc4f5d8c134262dd05caec464f642b8c585c119"} Nov 29 04:58:19 crc kubenswrapper[4799]: I1129 04:58:19.504873 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bfbbd39d-1e8f-4bdd-991e-71e206bb4ca0","Type":"ContainerStarted","Data":"7678ca66691ef3c869a15176a65df9215f29c89502a82713ab996e1993d11f45"} Nov 29 04:58:19 crc kubenswrapper[4799]: I1129 04:58:19.524703 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.524673455 podStartE2EDuration="2.524673455s" podCreationTimestamp="2025-11-29 04:58:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:58:19.521688474 +0000 UTC m=+1175.164618874" watchObservedRunningTime="2025-11-29 04:58:19.524673455 +0000 UTC m=+1175.167603855" Nov 29 04:58:22 crc kubenswrapper[4799]: I1129 04:58:22.184156 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 29 04:58:22 crc kubenswrapper[4799]: I1129 04:58:22.856903 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 29 04:58:22 crc kubenswrapper[4799]: I1129 04:58:22.856989 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 29 04:58:26 crc kubenswrapper[4799]: I1129 04:58:26.171388 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 29 04:58:26 crc kubenswrapper[4799]: I1129 04:58:26.171730 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 29 04:58:27 crc kubenswrapper[4799]: I1129 04:58:27.183969 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="cd1dc76e-76ba-4ab4-b40c-587ffbf9b5df" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.183:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 29 04:58:27 crc kubenswrapper[4799]: I1129 04:58:27.184119 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 29 04:58:27 crc kubenswrapper[4799]: I1129 04:58:27.184215 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="cd1dc76e-76ba-4ab4-b40c-587ffbf9b5df" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.183:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 29 04:58:27 crc kubenswrapper[4799]: I1129 04:58:27.217994 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 29 04:58:27 crc kubenswrapper[4799]: I1129 04:58:27.634156 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 29 04:58:27 crc kubenswrapper[4799]: I1129 04:58:27.856367 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 29 04:58:27 crc kubenswrapper[4799]: I1129 04:58:27.856427 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 29 04:58:28 crc kubenswrapper[4799]: I1129 04:58:28.716841 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 29 04:58:28 crc kubenswrapper[4799]: I1129 04:58:28.872032 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="bfbbd39d-1e8f-4bdd-991e-71e206bb4ca0" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.185:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 29 04:58:28 crc kubenswrapper[4799]: I1129 04:58:28.872860 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="bfbbd39d-1e8f-4bdd-991e-71e206bb4ca0" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.185:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 29 04:58:36 crc kubenswrapper[4799]: I1129 04:58:36.177860 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 29 04:58:36 crc kubenswrapper[4799]: I1129 04:58:36.178527 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 29 04:58:36 crc kubenswrapper[4799]: I1129 04:58:36.178917 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 29 04:58:36 crc kubenswrapper[4799]: I1129 04:58:36.178957 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 29 04:58:36 crc kubenswrapper[4799]: I1129 04:58:36.185479 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 29 04:58:36 crc kubenswrapper[4799]: I1129 04:58:36.186921 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 29 04:58:37 crc kubenswrapper[4799]: I1129 04:58:37.497177 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 04:58:37 crc kubenswrapper[4799]: I1129 04:58:37.497500 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 04:58:37 crc kubenswrapper[4799]: I1129 04:58:37.497554 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-26t88" Nov 29 04:58:37 crc kubenswrapper[4799]: I1129 04:58:37.498390 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0bb5e3708c6fb82d15eb8dc7170b3147f6cdc27661c141f7e521f399422452a4"} pod="openshift-machine-config-operator/machine-config-daemon-26t88" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 29 04:58:37 crc kubenswrapper[4799]: I1129 04:58:37.498460 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" containerID="cri-o://0bb5e3708c6fb82d15eb8dc7170b3147f6cdc27661c141f7e521f399422452a4" gracePeriod=600 Nov 29 04:58:37 crc kubenswrapper[4799]: I1129 04:58:37.712439 4799 generic.go:334] "Generic (PLEG): container finished" podID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerID="0bb5e3708c6fb82d15eb8dc7170b3147f6cdc27661c141f7e521f399422452a4" exitCode=0 Nov 29 04:58:37 crc kubenswrapper[4799]: I1129 04:58:37.712529 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" event={"ID":"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8","Type":"ContainerDied","Data":"0bb5e3708c6fb82d15eb8dc7170b3147f6cdc27661c141f7e521f399422452a4"} Nov 29 04:58:37 crc kubenswrapper[4799]: I1129 04:58:37.712764 4799 scope.go:117] "RemoveContainer" containerID="8a07429ac3c8e7bd45ed97e1ba2d6c50f73c78cd13fcb21a144bfb0b7bc995f6" Nov 29 04:58:37 crc kubenswrapper[4799]: I1129 04:58:37.863002 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 29 04:58:37 crc kubenswrapper[4799]: I1129 04:58:37.863349 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 29 04:58:37 crc kubenswrapper[4799]: I1129 04:58:37.869276 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 29 04:58:37 crc kubenswrapper[4799]: I1129 04:58:37.874425 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 29 04:58:38 crc kubenswrapper[4799]: I1129 04:58:38.723504 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" event={"ID":"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8","Type":"ContainerStarted","Data":"1b0c3452acd5fda0514988bec8ce2870ba0271b8c8c2d6b63a0f1b06bb0db0a1"} Nov 29 04:58:46 crc kubenswrapper[4799]: I1129 04:58:46.243311 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 29 04:58:47 crc kubenswrapper[4799]: I1129 04:58:47.057709 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 29 04:58:50 crc kubenswrapper[4799]: I1129 04:58:50.446747 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="1788e4d3-8c2a-4d06-b746-c0fb12514148" containerName="rabbitmq" containerID="cri-o://7abd82f4581f37354e1f18a9c0d46bf467ffbb2b06cbe83edec2915ef5bd8a74" gracePeriod=604796 Nov 29 04:58:51 crc kubenswrapper[4799]: I1129 04:58:51.185500 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="81726039-72e9-455a-b4fb-e16184acdc47" containerName="rabbitmq" containerID="cri-o://037ee3bbc021dfbf6239503e300c61dac479169025f6c1c3911be93913381a66" gracePeriod=604796 Nov 29 04:58:51 crc kubenswrapper[4799]: I1129 04:58:51.766724 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="1788e4d3-8c2a-4d06-b746-c0fb12514148" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.94:5671: connect: connection refused" Nov 29 04:58:52 crc kubenswrapper[4799]: I1129 04:58:52.111213 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="81726039-72e9-455a-b4fb-e16184acdc47" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.95:5671: connect: connection refused" Nov 29 04:58:56 crc kubenswrapper[4799]: I1129 04:58:56.880358 4799 generic.go:334] "Generic (PLEG): container finished" podID="1788e4d3-8c2a-4d06-b746-c0fb12514148" containerID="7abd82f4581f37354e1f18a9c0d46bf467ffbb2b06cbe83edec2915ef5bd8a74" exitCode=0 Nov 29 04:58:56 crc kubenswrapper[4799]: I1129 04:58:56.880425 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1788e4d3-8c2a-4d06-b746-c0fb12514148","Type":"ContainerDied","Data":"7abd82f4581f37354e1f18a9c0d46bf467ffbb2b06cbe83edec2915ef5bd8a74"} Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.127286 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.277064 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1788e4d3-8c2a-4d06-b746-c0fb12514148-server-conf\") pod \"1788e4d3-8c2a-4d06-b746-c0fb12514148\" (UID: \"1788e4d3-8c2a-4d06-b746-c0fb12514148\") " Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.277106 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1788e4d3-8c2a-4d06-b746-c0fb12514148-pod-info\") pod \"1788e4d3-8c2a-4d06-b746-c0fb12514148\" (UID: \"1788e4d3-8c2a-4d06-b746-c0fb12514148\") " Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.277157 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1788e4d3-8c2a-4d06-b746-c0fb12514148-rabbitmq-erlang-cookie\") pod \"1788e4d3-8c2a-4d06-b746-c0fb12514148\" (UID: \"1788e4d3-8c2a-4d06-b746-c0fb12514148\") " Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.277191 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1788e4d3-8c2a-4d06-b746-c0fb12514148-plugins-conf\") pod \"1788e4d3-8c2a-4d06-b746-c0fb12514148\" (UID: \"1788e4d3-8c2a-4d06-b746-c0fb12514148\") " Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.277229 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k94v6\" (UniqueName: \"kubernetes.io/projected/1788e4d3-8c2a-4d06-b746-c0fb12514148-kube-api-access-k94v6\") pod \"1788e4d3-8c2a-4d06-b746-c0fb12514148\" (UID: \"1788e4d3-8c2a-4d06-b746-c0fb12514148\") " Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.277257 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1788e4d3-8c2a-4d06-b746-c0fb12514148-rabbitmq-confd\") pod \"1788e4d3-8c2a-4d06-b746-c0fb12514148\" (UID: \"1788e4d3-8c2a-4d06-b746-c0fb12514148\") " Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.277281 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"1788e4d3-8c2a-4d06-b746-c0fb12514148\" (UID: \"1788e4d3-8c2a-4d06-b746-c0fb12514148\") " Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.277353 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1788e4d3-8c2a-4d06-b746-c0fb12514148-rabbitmq-plugins\") pod \"1788e4d3-8c2a-4d06-b746-c0fb12514148\" (UID: \"1788e4d3-8c2a-4d06-b746-c0fb12514148\") " Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.277391 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1788e4d3-8c2a-4d06-b746-c0fb12514148-config-data\") pod \"1788e4d3-8c2a-4d06-b746-c0fb12514148\" (UID: \"1788e4d3-8c2a-4d06-b746-c0fb12514148\") " Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.277445 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1788e4d3-8c2a-4d06-b746-c0fb12514148-erlang-cookie-secret\") pod \"1788e4d3-8c2a-4d06-b746-c0fb12514148\" (UID: \"1788e4d3-8c2a-4d06-b746-c0fb12514148\") " Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.277466 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1788e4d3-8c2a-4d06-b746-c0fb12514148-rabbitmq-tls\") pod \"1788e4d3-8c2a-4d06-b746-c0fb12514148\" (UID: \"1788e4d3-8c2a-4d06-b746-c0fb12514148\") " Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.279466 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1788e4d3-8c2a-4d06-b746-c0fb12514148-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "1788e4d3-8c2a-4d06-b746-c0fb12514148" (UID: "1788e4d3-8c2a-4d06-b746-c0fb12514148"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.279834 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1788e4d3-8c2a-4d06-b746-c0fb12514148-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "1788e4d3-8c2a-4d06-b746-c0fb12514148" (UID: "1788e4d3-8c2a-4d06-b746-c0fb12514148"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.281097 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1788e4d3-8c2a-4d06-b746-c0fb12514148-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "1788e4d3-8c2a-4d06-b746-c0fb12514148" (UID: "1788e4d3-8c2a-4d06-b746-c0fb12514148"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.284814 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "persistence") pod "1788e4d3-8c2a-4d06-b746-c0fb12514148" (UID: "1788e4d3-8c2a-4d06-b746-c0fb12514148"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.285375 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/1788e4d3-8c2a-4d06-b746-c0fb12514148-pod-info" (OuterVolumeSpecName: "pod-info") pod "1788e4d3-8c2a-4d06-b746-c0fb12514148" (UID: "1788e4d3-8c2a-4d06-b746-c0fb12514148"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.285939 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1788e4d3-8c2a-4d06-b746-c0fb12514148-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "1788e4d3-8c2a-4d06-b746-c0fb12514148" (UID: "1788e4d3-8c2a-4d06-b746-c0fb12514148"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.289161 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1788e4d3-8c2a-4d06-b746-c0fb12514148-kube-api-access-k94v6" (OuterVolumeSpecName: "kube-api-access-k94v6") pod "1788e4d3-8c2a-4d06-b746-c0fb12514148" (UID: "1788e4d3-8c2a-4d06-b746-c0fb12514148"). InnerVolumeSpecName "kube-api-access-k94v6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.302914 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1788e4d3-8c2a-4d06-b746-c0fb12514148-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "1788e4d3-8c2a-4d06-b746-c0fb12514148" (UID: "1788e4d3-8c2a-4d06-b746-c0fb12514148"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.341743 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1788e4d3-8c2a-4d06-b746-c0fb12514148-config-data" (OuterVolumeSpecName: "config-data") pod "1788e4d3-8c2a-4d06-b746-c0fb12514148" (UID: "1788e4d3-8c2a-4d06-b746-c0fb12514148"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.364293 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1788e4d3-8c2a-4d06-b746-c0fb12514148-server-conf" (OuterVolumeSpecName: "server-conf") pod "1788e4d3-8c2a-4d06-b746-c0fb12514148" (UID: "1788e4d3-8c2a-4d06-b746-c0fb12514148"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.379922 4799 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1788e4d3-8c2a-4d06-b746-c0fb12514148-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.379976 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1788e4d3-8c2a-4d06-b746-c0fb12514148-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.380006 4799 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1788e4d3-8c2a-4d06-b746-c0fb12514148-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.380019 4799 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1788e4d3-8c2a-4d06-b746-c0fb12514148-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.380073 4799 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1788e4d3-8c2a-4d06-b746-c0fb12514148-server-conf\") on node \"crc\" DevicePath \"\"" Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.380084 4799 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1788e4d3-8c2a-4d06-b746-c0fb12514148-pod-info\") on node \"crc\" DevicePath \"\"" Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.380095 4799 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1788e4d3-8c2a-4d06-b746-c0fb12514148-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.380109 4799 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1788e4d3-8c2a-4d06-b746-c0fb12514148-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.380119 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k94v6\" (UniqueName: \"kubernetes.io/projected/1788e4d3-8c2a-4d06-b746-c0fb12514148-kube-api-access-k94v6\") on node \"crc\" DevicePath \"\"" Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.380154 4799 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.408780 4799 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.415460 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1788e4d3-8c2a-4d06-b746-c0fb12514148-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "1788e4d3-8c2a-4d06-b746-c0fb12514148" (UID: "1788e4d3-8c2a-4d06-b746-c0fb12514148"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.482105 4799 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1788e4d3-8c2a-4d06-b746-c0fb12514148-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.482397 4799 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.881758 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.888684 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/81726039-72e9-455a-b4fb-e16184acdc47-rabbitmq-tls\") pod \"81726039-72e9-455a-b4fb-e16184acdc47\" (UID: \"81726039-72e9-455a-b4fb-e16184acdc47\") " Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.888731 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"81726039-72e9-455a-b4fb-e16184acdc47\" (UID: \"81726039-72e9-455a-b4fb-e16184acdc47\") " Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.888813 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/81726039-72e9-455a-b4fb-e16184acdc47-rabbitmq-erlang-cookie\") pod \"81726039-72e9-455a-b4fb-e16184acdc47\" (UID: \"81726039-72e9-455a-b4fb-e16184acdc47\") " Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.888838 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/81726039-72e9-455a-b4fb-e16184acdc47-server-conf\") pod \"81726039-72e9-455a-b4fb-e16184acdc47\" (UID: \"81726039-72e9-455a-b4fb-e16184acdc47\") " Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.888881 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/81726039-72e9-455a-b4fb-e16184acdc47-erlang-cookie-secret\") pod \"81726039-72e9-455a-b4fb-e16184acdc47\" (UID: \"81726039-72e9-455a-b4fb-e16184acdc47\") " Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.888913 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/81726039-72e9-455a-b4fb-e16184acdc47-rabbitmq-confd\") pod \"81726039-72e9-455a-b4fb-e16184acdc47\" (UID: \"81726039-72e9-455a-b4fb-e16184acdc47\") " Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.888935 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z859g\" (UniqueName: \"kubernetes.io/projected/81726039-72e9-455a-b4fb-e16184acdc47-kube-api-access-z859g\") pod \"81726039-72e9-455a-b4fb-e16184acdc47\" (UID: \"81726039-72e9-455a-b4fb-e16184acdc47\") " Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.888951 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/81726039-72e9-455a-b4fb-e16184acdc47-config-data\") pod \"81726039-72e9-455a-b4fb-e16184acdc47\" (UID: \"81726039-72e9-455a-b4fb-e16184acdc47\") " Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.888977 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/81726039-72e9-455a-b4fb-e16184acdc47-pod-info\") pod \"81726039-72e9-455a-b4fb-e16184acdc47\" (UID: \"81726039-72e9-455a-b4fb-e16184acdc47\") " Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.888999 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/81726039-72e9-455a-b4fb-e16184acdc47-plugins-conf\") pod \"81726039-72e9-455a-b4fb-e16184acdc47\" (UID: \"81726039-72e9-455a-b4fb-e16184acdc47\") " Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.889035 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/81726039-72e9-455a-b4fb-e16184acdc47-rabbitmq-plugins\") pod \"81726039-72e9-455a-b4fb-e16184acdc47\" (UID: \"81726039-72e9-455a-b4fb-e16184acdc47\") " Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.889373 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81726039-72e9-455a-b4fb-e16184acdc47-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "81726039-72e9-455a-b4fb-e16184acdc47" (UID: "81726039-72e9-455a-b4fb-e16184acdc47"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.889837 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81726039-72e9-455a-b4fb-e16184acdc47-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "81726039-72e9-455a-b4fb-e16184acdc47" (UID: "81726039-72e9-455a-b4fb-e16184acdc47"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.890294 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81726039-72e9-455a-b4fb-e16184acdc47-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "81726039-72e9-455a-b4fb-e16184acdc47" (UID: "81726039-72e9-455a-b4fb-e16184acdc47"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.893713 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/81726039-72e9-455a-b4fb-e16184acdc47-pod-info" (OuterVolumeSpecName: "pod-info") pod "81726039-72e9-455a-b4fb-e16184acdc47" (UID: "81726039-72e9-455a-b4fb-e16184acdc47"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.893763 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81726039-72e9-455a-b4fb-e16184acdc47-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "81726039-72e9-455a-b4fb-e16184acdc47" (UID: "81726039-72e9-455a-b4fb-e16184acdc47"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.895199 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1788e4d3-8c2a-4d06-b746-c0fb12514148","Type":"ContainerDied","Data":"25b331b3efc2702cb80297c00783deea910cc25e6c427e635005bcaff6f39e61"} Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.895240 4799 scope.go:117] "RemoveContainer" containerID="7abd82f4581f37354e1f18a9c0d46bf467ffbb2b06cbe83edec2915ef5bd8a74" Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.895354 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.897142 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81726039-72e9-455a-b4fb-e16184acdc47-kube-api-access-z859g" (OuterVolumeSpecName: "kube-api-access-z859g") pod "81726039-72e9-455a-b4fb-e16184acdc47" (UID: "81726039-72e9-455a-b4fb-e16184acdc47"). InnerVolumeSpecName "kube-api-access-z859g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.902137 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "persistence") pod "81726039-72e9-455a-b4fb-e16184acdc47" (UID: "81726039-72e9-455a-b4fb-e16184acdc47"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.902953 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.902781 4799 generic.go:334] "Generic (PLEG): container finished" podID="81726039-72e9-455a-b4fb-e16184acdc47" containerID="037ee3bbc021dfbf6239503e300c61dac479169025f6c1c3911be93913381a66" exitCode=0 Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.903185 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"81726039-72e9-455a-b4fb-e16184acdc47","Type":"ContainerDied","Data":"037ee3bbc021dfbf6239503e300c61dac479169025f6c1c3911be93913381a66"} Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.903585 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"81726039-72e9-455a-b4fb-e16184acdc47","Type":"ContainerDied","Data":"3bfd2d2b4232ba9982c50d11901772e1aed4dcbdae556a40673d8ea38d646aa1"} Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.904492 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81726039-72e9-455a-b4fb-e16184acdc47-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "81726039-72e9-455a-b4fb-e16184acdc47" (UID: "81726039-72e9-455a-b4fb-e16184acdc47"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:58:57 crc kubenswrapper[4799]: I1129 04:58:57.979226 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81726039-72e9-455a-b4fb-e16184acdc47-config-data" (OuterVolumeSpecName: "config-data") pod "81726039-72e9-455a-b4fb-e16184acdc47" (UID: "81726039-72e9-455a-b4fb-e16184acdc47"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.001317 4799 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/81726039-72e9-455a-b4fb-e16184acdc47-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.003750 4799 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/81726039-72e9-455a-b4fb-e16184acdc47-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.008184 4799 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/81726039-72e9-455a-b4fb-e16184acdc47-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.008365 4799 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.008449 4799 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/81726039-72e9-455a-b4fb-e16184acdc47-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.008527 4799 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/81726039-72e9-455a-b4fb-e16184acdc47-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.008605 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z859g\" (UniqueName: \"kubernetes.io/projected/81726039-72e9-455a-b4fb-e16184acdc47-kube-api-access-z859g\") on node \"crc\" DevicePath \"\"" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.008680 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/81726039-72e9-455a-b4fb-e16184acdc47-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.008756 4799 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/81726039-72e9-455a-b4fb-e16184acdc47-pod-info\") on node \"crc\" DevicePath \"\"" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.023671 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81726039-72e9-455a-b4fb-e16184acdc47-server-conf" (OuterVolumeSpecName: "server-conf") pod "81726039-72e9-455a-b4fb-e16184acdc47" (UID: "81726039-72e9-455a-b4fb-e16184acdc47"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.057032 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81726039-72e9-455a-b4fb-e16184acdc47-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "81726039-72e9-455a-b4fb-e16184acdc47" (UID: "81726039-72e9-455a-b4fb-e16184acdc47"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.065855 4799 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.105064 4799 scope.go:117] "RemoveContainer" containerID="c7803748c9deb308ab6a3661ce9b6ea289206d8f547aa0262d26c7a7a46628cc" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.112832 4799 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/81726039-72e9-455a-b4fb-e16184acdc47-server-conf\") on node \"crc\" DevicePath \"\"" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.112913 4799 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/81726039-72e9-455a-b4fb-e16184acdc47-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.112932 4799 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.147824 4799 scope.go:117] "RemoveContainer" containerID="037ee3bbc021dfbf6239503e300c61dac479169025f6c1c3911be93913381a66" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.164871 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.175855 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.185737 4799 scope.go:117] "RemoveContainer" containerID="a72d02dc2cf8aa7523907bde01f61568e9db9f50c267ea96039aa2db0a285b08" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.187028 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 29 04:58:58 crc kubenswrapper[4799]: E1129 04:58:58.187357 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81726039-72e9-455a-b4fb-e16184acdc47" containerName="setup-container" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.187375 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="81726039-72e9-455a-b4fb-e16184acdc47" containerName="setup-container" Nov 29 04:58:58 crc kubenswrapper[4799]: E1129 04:58:58.187402 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1788e4d3-8c2a-4d06-b746-c0fb12514148" containerName="rabbitmq" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.187408 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="1788e4d3-8c2a-4d06-b746-c0fb12514148" containerName="rabbitmq" Nov 29 04:58:58 crc kubenswrapper[4799]: E1129 04:58:58.187423 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81726039-72e9-455a-b4fb-e16184acdc47" containerName="rabbitmq" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.187428 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="81726039-72e9-455a-b4fb-e16184acdc47" containerName="rabbitmq" Nov 29 04:58:58 crc kubenswrapper[4799]: E1129 04:58:58.187437 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1788e4d3-8c2a-4d06-b746-c0fb12514148" containerName="setup-container" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.187443 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="1788e4d3-8c2a-4d06-b746-c0fb12514148" containerName="setup-container" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.187629 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="81726039-72e9-455a-b4fb-e16184acdc47" containerName="rabbitmq" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.187659 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="1788e4d3-8c2a-4d06-b746-c0fb12514148" containerName="rabbitmq" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.188812 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.191629 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.191892 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.192226 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.193585 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.193672 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.193834 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.194324 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-7gp88" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.200051 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.218188 4799 scope.go:117] "RemoveContainer" containerID="037ee3bbc021dfbf6239503e300c61dac479169025f6c1c3911be93913381a66" Nov 29 04:58:58 crc kubenswrapper[4799]: E1129 04:58:58.220218 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"037ee3bbc021dfbf6239503e300c61dac479169025f6c1c3911be93913381a66\": container with ID starting with 037ee3bbc021dfbf6239503e300c61dac479169025f6c1c3911be93913381a66 not found: ID does not exist" containerID="037ee3bbc021dfbf6239503e300c61dac479169025f6c1c3911be93913381a66" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.220297 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"037ee3bbc021dfbf6239503e300c61dac479169025f6c1c3911be93913381a66"} err="failed to get container status \"037ee3bbc021dfbf6239503e300c61dac479169025f6c1c3911be93913381a66\": rpc error: code = NotFound desc = could not find container \"037ee3bbc021dfbf6239503e300c61dac479169025f6c1c3911be93913381a66\": container with ID starting with 037ee3bbc021dfbf6239503e300c61dac479169025f6c1c3911be93913381a66 not found: ID does not exist" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.220464 4799 scope.go:117] "RemoveContainer" containerID="a72d02dc2cf8aa7523907bde01f61568e9db9f50c267ea96039aa2db0a285b08" Nov 29 04:58:58 crc kubenswrapper[4799]: E1129 04:58:58.221072 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a72d02dc2cf8aa7523907bde01f61568e9db9f50c267ea96039aa2db0a285b08\": container with ID starting with a72d02dc2cf8aa7523907bde01f61568e9db9f50c267ea96039aa2db0a285b08 not found: ID does not exist" containerID="a72d02dc2cf8aa7523907bde01f61568e9db9f50c267ea96039aa2db0a285b08" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.221158 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a72d02dc2cf8aa7523907bde01f61568e9db9f50c267ea96039aa2db0a285b08"} err="failed to get container status \"a72d02dc2cf8aa7523907bde01f61568e9db9f50c267ea96039aa2db0a285b08\": rpc error: code = NotFound desc = could not find container \"a72d02dc2cf8aa7523907bde01f61568e9db9f50c267ea96039aa2db0a285b08\": container with ID starting with a72d02dc2cf8aa7523907bde01f61568e9db9f50c267ea96039aa2db0a285b08 not found: ID does not exist" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.282881 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.296885 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.307822 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.309240 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.314661 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.314920 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.315098 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.315343 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.315558 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.315699 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.315918 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-jhdhp" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.318392 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0bfecf67-24fe-42fa-bad5-4a597ad9d3af-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0bfecf67-24fe-42fa-bad5-4a597ad9d3af\") " pod="openstack/rabbitmq-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.318431 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"0bfecf67-24fe-42fa-bad5-4a597ad9d3af\") " pod="openstack/rabbitmq-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.318480 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0bfecf67-24fe-42fa-bad5-4a597ad9d3af-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0bfecf67-24fe-42fa-bad5-4a597ad9d3af\") " pod="openstack/rabbitmq-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.318512 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0bfecf67-24fe-42fa-bad5-4a597ad9d3af-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0bfecf67-24fe-42fa-bad5-4a597ad9d3af\") " pod="openstack/rabbitmq-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.318556 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0bfecf67-24fe-42fa-bad5-4a597ad9d3af-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0bfecf67-24fe-42fa-bad5-4a597ad9d3af\") " pod="openstack/rabbitmq-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.318608 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zw5sx\" (UniqueName: \"kubernetes.io/projected/0bfecf67-24fe-42fa-bad5-4a597ad9d3af-kube-api-access-zw5sx\") pod \"rabbitmq-server-0\" (UID: \"0bfecf67-24fe-42fa-bad5-4a597ad9d3af\") " pod="openstack/rabbitmq-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.318644 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0bfecf67-24fe-42fa-bad5-4a597ad9d3af-config-data\") pod \"rabbitmq-server-0\" (UID: \"0bfecf67-24fe-42fa-bad5-4a597ad9d3af\") " pod="openstack/rabbitmq-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.318675 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0bfecf67-24fe-42fa-bad5-4a597ad9d3af-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"0bfecf67-24fe-42fa-bad5-4a597ad9d3af\") " pod="openstack/rabbitmq-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.318729 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0bfecf67-24fe-42fa-bad5-4a597ad9d3af-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0bfecf67-24fe-42fa-bad5-4a597ad9d3af\") " pod="openstack/rabbitmq-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.318756 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0bfecf67-24fe-42fa-bad5-4a597ad9d3af-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0bfecf67-24fe-42fa-bad5-4a597ad9d3af\") " pod="openstack/rabbitmq-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.318812 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0bfecf67-24fe-42fa-bad5-4a597ad9d3af-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0bfecf67-24fe-42fa-bad5-4a597ad9d3af\") " pod="openstack/rabbitmq-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.345727 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.420245 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0bfecf67-24fe-42fa-bad5-4a597ad9d3af-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0bfecf67-24fe-42fa-bad5-4a597ad9d3af\") " pod="openstack/rabbitmq-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.420296 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"0bfecf67-24fe-42fa-bad5-4a597ad9d3af\") " pod="openstack/rabbitmq-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.420343 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0bfecf67-24fe-42fa-bad5-4a597ad9d3af-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0bfecf67-24fe-42fa-bad5-4a597ad9d3af\") " pod="openstack/rabbitmq-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.420379 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0bfecf67-24fe-42fa-bad5-4a597ad9d3af-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0bfecf67-24fe-42fa-bad5-4a597ad9d3af\") " pod="openstack/rabbitmq-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.420411 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.421516 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.421066 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0bfecf67-24fe-42fa-bad5-4a597ad9d3af-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0bfecf67-24fe-42fa-bad5-4a597ad9d3af\") " pod="openstack/rabbitmq-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.421435 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0bfecf67-24fe-42fa-bad5-4a597ad9d3af-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0bfecf67-24fe-42fa-bad5-4a597ad9d3af\") " pod="openstack/rabbitmq-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.420625 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"0bfecf67-24fe-42fa-bad5-4a597ad9d3af\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.421590 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.421632 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0bfecf67-24fe-42fa-bad5-4a597ad9d3af-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0bfecf67-24fe-42fa-bad5-4a597ad9d3af\") " pod="openstack/rabbitmq-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.435825 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0bfecf67-24fe-42fa-bad5-4a597ad9d3af-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0bfecf67-24fe-42fa-bad5-4a597ad9d3af\") " pod="openstack/rabbitmq-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.435924 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.436176 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.436614 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zw5sx\" (UniqueName: \"kubernetes.io/projected/0bfecf67-24fe-42fa-bad5-4a597ad9d3af-kube-api-access-zw5sx\") pod \"rabbitmq-server-0\" (UID: \"0bfecf67-24fe-42fa-bad5-4a597ad9d3af\") " pod="openstack/rabbitmq-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.436862 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.436963 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0bfecf67-24fe-42fa-bad5-4a597ad9d3af-config-data\") pod \"rabbitmq-server-0\" (UID: \"0bfecf67-24fe-42fa-bad5-4a597ad9d3af\") " pod="openstack/rabbitmq-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.437022 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.437052 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0bfecf67-24fe-42fa-bad5-4a597ad9d3af-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"0bfecf67-24fe-42fa-bad5-4a597ad9d3af\") " pod="openstack/rabbitmq-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.437183 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0bfecf67-24fe-42fa-bad5-4a597ad9d3af-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0bfecf67-24fe-42fa-bad5-4a597ad9d3af\") " pod="openstack/rabbitmq-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.437238 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0bfecf67-24fe-42fa-bad5-4a597ad9d3af-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0bfecf67-24fe-42fa-bad5-4a597ad9d3af\") " pod="openstack/rabbitmq-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.437275 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4t9cv\" (UniqueName: \"kubernetes.io/projected/1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac-kube-api-access-4t9cv\") pod \"rabbitmq-cell1-server-0\" (UID: \"1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.437320 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0bfecf67-24fe-42fa-bad5-4a597ad9d3af-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0bfecf67-24fe-42fa-bad5-4a597ad9d3af\") " pod="openstack/rabbitmq-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.437352 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.437406 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.437436 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.439241 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0bfecf67-24fe-42fa-bad5-4a597ad9d3af-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0bfecf67-24fe-42fa-bad5-4a597ad9d3af\") " pod="openstack/rabbitmq-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.439434 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0bfecf67-24fe-42fa-bad5-4a597ad9d3af-config-data\") pod \"rabbitmq-server-0\" (UID: \"0bfecf67-24fe-42fa-bad5-4a597ad9d3af\") " pod="openstack/rabbitmq-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.451573 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0bfecf67-24fe-42fa-bad5-4a597ad9d3af-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0bfecf67-24fe-42fa-bad5-4a597ad9d3af\") " pod="openstack/rabbitmq-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.452651 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0bfecf67-24fe-42fa-bad5-4a597ad9d3af-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"0bfecf67-24fe-42fa-bad5-4a597ad9d3af\") " pod="openstack/rabbitmq-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.453463 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0bfecf67-24fe-42fa-bad5-4a597ad9d3af-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0bfecf67-24fe-42fa-bad5-4a597ad9d3af\") " pod="openstack/rabbitmq-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.453671 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0bfecf67-24fe-42fa-bad5-4a597ad9d3af-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0bfecf67-24fe-42fa-bad5-4a597ad9d3af\") " pod="openstack/rabbitmq-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.466980 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zw5sx\" (UniqueName: \"kubernetes.io/projected/0bfecf67-24fe-42fa-bad5-4a597ad9d3af-kube-api-access-zw5sx\") pod \"rabbitmq-server-0\" (UID: \"0bfecf67-24fe-42fa-bad5-4a597ad9d3af\") " pod="openstack/rabbitmq-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.468829 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"0bfecf67-24fe-42fa-bad5-4a597ad9d3af\") " pod="openstack/rabbitmq-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.517772 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.540250 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.540359 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.540397 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.540486 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.540522 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.540554 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.540597 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.540633 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.540678 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.540729 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.540831 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4t9cv\" (UniqueName: \"kubernetes.io/projected/1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac-kube-api-access-4t9cv\") pod \"rabbitmq-cell1-server-0\" (UID: \"1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.540856 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.541168 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.542137 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.542207 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.542840 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.543074 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.544250 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.548772 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.548845 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.549224 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.564621 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4t9cv\" (UniqueName: \"kubernetes.io/projected/1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac-kube-api-access-4t9cv\") pod \"rabbitmq-cell1-server-0\" (UID: \"1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.586907 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.634305 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.675466 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1788e4d3-8c2a-4d06-b746-c0fb12514148" path="/var/lib/kubelet/pods/1788e4d3-8c2a-4d06-b746-c0fb12514148/volumes" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.676618 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81726039-72e9-455a-b4fb-e16184acdc47" path="/var/lib/kubelet/pods/81726039-72e9-455a-b4fb-e16184acdc47/volumes" Nov 29 04:58:58 crc kubenswrapper[4799]: I1129 04:58:58.976702 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 29 04:58:59 crc kubenswrapper[4799]: I1129 04:58:59.145381 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 29 04:58:59 crc kubenswrapper[4799]: W1129 04:58:59.150153 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1f5c1a3b_7a4e_48df_aaff_de72acf9b1ac.slice/crio-25b40500f4f20dfe4445d1d705dca1ad3512cd5775cdf21699195533be0a5309 WatchSource:0}: Error finding container 25b40500f4f20dfe4445d1d705dca1ad3512cd5775cdf21699195533be0a5309: Status 404 returned error can't find the container with id 25b40500f4f20dfe4445d1d705dca1ad3512cd5775cdf21699195533be0a5309 Nov 29 04:58:59 crc kubenswrapper[4799]: I1129 04:58:59.928275 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac","Type":"ContainerStarted","Data":"25b40500f4f20dfe4445d1d705dca1ad3512cd5775cdf21699195533be0a5309"} Nov 29 04:58:59 crc kubenswrapper[4799]: I1129 04:58:59.929903 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0bfecf67-24fe-42fa-bad5-4a597ad9d3af","Type":"ContainerStarted","Data":"45f59ab4da9de09b8bbd6cfc1eedfa32e74aff3ebed930c251959a2093fc7aad"} Nov 29 04:59:00 crc kubenswrapper[4799]: I1129 04:59:00.938191 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac","Type":"ContainerStarted","Data":"abe77cc4d36024c6b591d4d16586151c95686b7404c44a477d7d08d78de919ee"} Nov 29 04:59:00 crc kubenswrapper[4799]: I1129 04:59:00.940669 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0bfecf67-24fe-42fa-bad5-4a597ad9d3af","Type":"ContainerStarted","Data":"c3f70cf40e8091597499a0c128f7230a1c8d1dca9ef99085ae1033299fe6d684"} Nov 29 04:59:01 crc kubenswrapper[4799]: I1129 04:59:01.167172 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6447ccbd8f-cvd9n"] Nov 29 04:59:01 crc kubenswrapper[4799]: I1129 04:59:01.169224 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6447ccbd8f-cvd9n" Nov 29 04:59:01 crc kubenswrapper[4799]: I1129 04:59:01.171881 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Nov 29 04:59:01 crc kubenswrapper[4799]: I1129 04:59:01.179249 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6447ccbd8f-cvd9n"] Nov 29 04:59:01 crc kubenswrapper[4799]: I1129 04:59:01.296521 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a332883-d0b4-483f-8ae9-93cf97d2815a-config\") pod \"dnsmasq-dns-6447ccbd8f-cvd9n\" (UID: \"0a332883-d0b4-483f-8ae9-93cf97d2815a\") " pod="openstack/dnsmasq-dns-6447ccbd8f-cvd9n" Nov 29 04:59:01 crc kubenswrapper[4799]: I1129 04:59:01.296592 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0a332883-d0b4-483f-8ae9-93cf97d2815a-openstack-edpm-ipam\") pod \"dnsmasq-dns-6447ccbd8f-cvd9n\" (UID: \"0a332883-d0b4-483f-8ae9-93cf97d2815a\") " pod="openstack/dnsmasq-dns-6447ccbd8f-cvd9n" Nov 29 04:59:01 crc kubenswrapper[4799]: I1129 04:59:01.296645 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5djw\" (UniqueName: \"kubernetes.io/projected/0a332883-d0b4-483f-8ae9-93cf97d2815a-kube-api-access-k5djw\") pod \"dnsmasq-dns-6447ccbd8f-cvd9n\" (UID: \"0a332883-d0b4-483f-8ae9-93cf97d2815a\") " pod="openstack/dnsmasq-dns-6447ccbd8f-cvd9n" Nov 29 04:59:01 crc kubenswrapper[4799]: I1129 04:59:01.296712 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0a332883-d0b4-483f-8ae9-93cf97d2815a-ovsdbserver-nb\") pod \"dnsmasq-dns-6447ccbd8f-cvd9n\" (UID: \"0a332883-d0b4-483f-8ae9-93cf97d2815a\") " pod="openstack/dnsmasq-dns-6447ccbd8f-cvd9n" Nov 29 04:59:01 crc kubenswrapper[4799]: I1129 04:59:01.296837 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0a332883-d0b4-483f-8ae9-93cf97d2815a-ovsdbserver-sb\") pod \"dnsmasq-dns-6447ccbd8f-cvd9n\" (UID: \"0a332883-d0b4-483f-8ae9-93cf97d2815a\") " pod="openstack/dnsmasq-dns-6447ccbd8f-cvd9n" Nov 29 04:59:01 crc kubenswrapper[4799]: I1129 04:59:01.296866 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a332883-d0b4-483f-8ae9-93cf97d2815a-dns-svc\") pod \"dnsmasq-dns-6447ccbd8f-cvd9n\" (UID: \"0a332883-d0b4-483f-8ae9-93cf97d2815a\") " pod="openstack/dnsmasq-dns-6447ccbd8f-cvd9n" Nov 29 04:59:01 crc kubenswrapper[4799]: I1129 04:59:01.398640 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0a332883-d0b4-483f-8ae9-93cf97d2815a-openstack-edpm-ipam\") pod \"dnsmasq-dns-6447ccbd8f-cvd9n\" (UID: \"0a332883-d0b4-483f-8ae9-93cf97d2815a\") " pod="openstack/dnsmasq-dns-6447ccbd8f-cvd9n" Nov 29 04:59:01 crc kubenswrapper[4799]: I1129 04:59:01.399007 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5djw\" (UniqueName: \"kubernetes.io/projected/0a332883-d0b4-483f-8ae9-93cf97d2815a-kube-api-access-k5djw\") pod \"dnsmasq-dns-6447ccbd8f-cvd9n\" (UID: \"0a332883-d0b4-483f-8ae9-93cf97d2815a\") " pod="openstack/dnsmasq-dns-6447ccbd8f-cvd9n" Nov 29 04:59:01 crc kubenswrapper[4799]: I1129 04:59:01.399055 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0a332883-d0b4-483f-8ae9-93cf97d2815a-ovsdbserver-nb\") pod \"dnsmasq-dns-6447ccbd8f-cvd9n\" (UID: \"0a332883-d0b4-483f-8ae9-93cf97d2815a\") " pod="openstack/dnsmasq-dns-6447ccbd8f-cvd9n" Nov 29 04:59:01 crc kubenswrapper[4799]: I1129 04:59:01.399119 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0a332883-d0b4-483f-8ae9-93cf97d2815a-ovsdbserver-sb\") pod \"dnsmasq-dns-6447ccbd8f-cvd9n\" (UID: \"0a332883-d0b4-483f-8ae9-93cf97d2815a\") " pod="openstack/dnsmasq-dns-6447ccbd8f-cvd9n" Nov 29 04:59:01 crc kubenswrapper[4799]: I1129 04:59:01.399145 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a332883-d0b4-483f-8ae9-93cf97d2815a-dns-svc\") pod \"dnsmasq-dns-6447ccbd8f-cvd9n\" (UID: \"0a332883-d0b4-483f-8ae9-93cf97d2815a\") " pod="openstack/dnsmasq-dns-6447ccbd8f-cvd9n" Nov 29 04:59:01 crc kubenswrapper[4799]: I1129 04:59:01.399217 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a332883-d0b4-483f-8ae9-93cf97d2815a-config\") pod \"dnsmasq-dns-6447ccbd8f-cvd9n\" (UID: \"0a332883-d0b4-483f-8ae9-93cf97d2815a\") " pod="openstack/dnsmasq-dns-6447ccbd8f-cvd9n" Nov 29 04:59:01 crc kubenswrapper[4799]: I1129 04:59:01.399763 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0a332883-d0b4-483f-8ae9-93cf97d2815a-openstack-edpm-ipam\") pod \"dnsmasq-dns-6447ccbd8f-cvd9n\" (UID: \"0a332883-d0b4-483f-8ae9-93cf97d2815a\") " pod="openstack/dnsmasq-dns-6447ccbd8f-cvd9n" Nov 29 04:59:01 crc kubenswrapper[4799]: I1129 04:59:01.400038 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0a332883-d0b4-483f-8ae9-93cf97d2815a-ovsdbserver-sb\") pod \"dnsmasq-dns-6447ccbd8f-cvd9n\" (UID: \"0a332883-d0b4-483f-8ae9-93cf97d2815a\") " pod="openstack/dnsmasq-dns-6447ccbd8f-cvd9n" Nov 29 04:59:01 crc kubenswrapper[4799]: I1129 04:59:01.400435 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0a332883-d0b4-483f-8ae9-93cf97d2815a-ovsdbserver-nb\") pod \"dnsmasq-dns-6447ccbd8f-cvd9n\" (UID: \"0a332883-d0b4-483f-8ae9-93cf97d2815a\") " pod="openstack/dnsmasq-dns-6447ccbd8f-cvd9n" Nov 29 04:59:01 crc kubenswrapper[4799]: I1129 04:59:01.400475 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a332883-d0b4-483f-8ae9-93cf97d2815a-dns-svc\") pod \"dnsmasq-dns-6447ccbd8f-cvd9n\" (UID: \"0a332883-d0b4-483f-8ae9-93cf97d2815a\") " pod="openstack/dnsmasq-dns-6447ccbd8f-cvd9n" Nov 29 04:59:01 crc kubenswrapper[4799]: I1129 04:59:01.401773 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a332883-d0b4-483f-8ae9-93cf97d2815a-config\") pod \"dnsmasq-dns-6447ccbd8f-cvd9n\" (UID: \"0a332883-d0b4-483f-8ae9-93cf97d2815a\") " pod="openstack/dnsmasq-dns-6447ccbd8f-cvd9n" Nov 29 04:59:01 crc kubenswrapper[4799]: I1129 04:59:01.417682 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5djw\" (UniqueName: \"kubernetes.io/projected/0a332883-d0b4-483f-8ae9-93cf97d2815a-kube-api-access-k5djw\") pod \"dnsmasq-dns-6447ccbd8f-cvd9n\" (UID: \"0a332883-d0b4-483f-8ae9-93cf97d2815a\") " pod="openstack/dnsmasq-dns-6447ccbd8f-cvd9n" Nov 29 04:59:01 crc kubenswrapper[4799]: I1129 04:59:01.497563 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6447ccbd8f-cvd9n" Nov 29 04:59:01 crc kubenswrapper[4799]: I1129 04:59:01.939874 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6447ccbd8f-cvd9n"] Nov 29 04:59:01 crc kubenswrapper[4799]: W1129 04:59:01.953816 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0a332883_d0b4_483f_8ae9_93cf97d2815a.slice/crio-1612a74094cc52052e6752ca2cecdadaf14119c22173fa9f33e3a99833e28b8f WatchSource:0}: Error finding container 1612a74094cc52052e6752ca2cecdadaf14119c22173fa9f33e3a99833e28b8f: Status 404 returned error can't find the container with id 1612a74094cc52052e6752ca2cecdadaf14119c22173fa9f33e3a99833e28b8f Nov 29 04:59:02 crc kubenswrapper[4799]: I1129 04:59:02.958417 4799 generic.go:334] "Generic (PLEG): container finished" podID="0a332883-d0b4-483f-8ae9-93cf97d2815a" containerID="7d1a52c97518110d33c0ca415378f6e995179958cd21b2f67174b585b8097df2" exitCode=0 Nov 29 04:59:02 crc kubenswrapper[4799]: I1129 04:59:02.958476 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6447ccbd8f-cvd9n" event={"ID":"0a332883-d0b4-483f-8ae9-93cf97d2815a","Type":"ContainerDied","Data":"7d1a52c97518110d33c0ca415378f6e995179958cd21b2f67174b585b8097df2"} Nov 29 04:59:02 crc kubenswrapper[4799]: I1129 04:59:02.958685 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6447ccbd8f-cvd9n" event={"ID":"0a332883-d0b4-483f-8ae9-93cf97d2815a","Type":"ContainerStarted","Data":"1612a74094cc52052e6752ca2cecdadaf14119c22173fa9f33e3a99833e28b8f"} Nov 29 04:59:03 crc kubenswrapper[4799]: I1129 04:59:03.972837 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6447ccbd8f-cvd9n" event={"ID":"0a332883-d0b4-483f-8ae9-93cf97d2815a","Type":"ContainerStarted","Data":"eadd10961388f58d49c48365e68b1dbc5b6715e2bf945f0943d0cef0351a0261"} Nov 29 04:59:03 crc kubenswrapper[4799]: I1129 04:59:03.973310 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6447ccbd8f-cvd9n" Nov 29 04:59:04 crc kubenswrapper[4799]: I1129 04:59:04.001676 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6447ccbd8f-cvd9n" podStartSLOduration=3.001650528 podStartE2EDuration="3.001650528s" podCreationTimestamp="2025-11-29 04:59:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:59:03.99374156 +0000 UTC m=+1219.636671980" watchObservedRunningTime="2025-11-29 04:59:04.001650528 +0000 UTC m=+1219.644580928" Nov 29 04:59:11 crc kubenswrapper[4799]: I1129 04:59:11.498922 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6447ccbd8f-cvd9n" Nov 29 04:59:11 crc kubenswrapper[4799]: I1129 04:59:11.603649 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b856c5697-jgvwh"] Nov 29 04:59:11 crc kubenswrapper[4799]: I1129 04:59:11.603948 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b856c5697-jgvwh" podUID="c0b29191-5d29-44fa-9cde-c2604b5426bb" containerName="dnsmasq-dns" containerID="cri-o://e02ae0768cfb16467fe1cee900bb77a9e6b8885860e9ce2e46625c69d51479b2" gracePeriod=10 Nov 29 04:59:11 crc kubenswrapper[4799]: I1129 04:59:11.655307 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5b856c5697-jgvwh" podUID="c0b29191-5d29-44fa-9cde-c2604b5426bb" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.179:5353: connect: connection refused" Nov 29 04:59:11 crc kubenswrapper[4799]: I1129 04:59:11.778804 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-864d5fc68c-cws22"] Nov 29 04:59:11 crc kubenswrapper[4799]: I1129 04:59:11.827890 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-864d5fc68c-cws22" Nov 29 04:59:11 crc kubenswrapper[4799]: I1129 04:59:11.840250 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-864d5fc68c-cws22"] Nov 29 04:59:11 crc kubenswrapper[4799]: I1129 04:59:11.961772 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee60cfe8-e5d9-4e9a-9299-38a14710fcdd-dns-svc\") pod \"dnsmasq-dns-864d5fc68c-cws22\" (UID: \"ee60cfe8-e5d9-4e9a-9299-38a14710fcdd\") " pod="openstack/dnsmasq-dns-864d5fc68c-cws22" Nov 29 04:59:11 crc kubenswrapper[4799]: I1129 04:59:11.961934 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ee60cfe8-e5d9-4e9a-9299-38a14710fcdd-ovsdbserver-sb\") pod \"dnsmasq-dns-864d5fc68c-cws22\" (UID: \"ee60cfe8-e5d9-4e9a-9299-38a14710fcdd\") " pod="openstack/dnsmasq-dns-864d5fc68c-cws22" Nov 29 04:59:11 crc kubenswrapper[4799]: I1129 04:59:11.961962 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87595\" (UniqueName: \"kubernetes.io/projected/ee60cfe8-e5d9-4e9a-9299-38a14710fcdd-kube-api-access-87595\") pod \"dnsmasq-dns-864d5fc68c-cws22\" (UID: \"ee60cfe8-e5d9-4e9a-9299-38a14710fcdd\") " pod="openstack/dnsmasq-dns-864d5fc68c-cws22" Nov 29 04:59:11 crc kubenswrapper[4799]: I1129 04:59:11.961983 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ee60cfe8-e5d9-4e9a-9299-38a14710fcdd-ovsdbserver-nb\") pod \"dnsmasq-dns-864d5fc68c-cws22\" (UID: \"ee60cfe8-e5d9-4e9a-9299-38a14710fcdd\") " pod="openstack/dnsmasq-dns-864d5fc68c-cws22" Nov 29 04:59:11 crc kubenswrapper[4799]: I1129 04:59:11.962001 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ee60cfe8-e5d9-4e9a-9299-38a14710fcdd-openstack-edpm-ipam\") pod \"dnsmasq-dns-864d5fc68c-cws22\" (UID: \"ee60cfe8-e5d9-4e9a-9299-38a14710fcdd\") " pod="openstack/dnsmasq-dns-864d5fc68c-cws22" Nov 29 04:59:11 crc kubenswrapper[4799]: I1129 04:59:11.962055 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee60cfe8-e5d9-4e9a-9299-38a14710fcdd-config\") pod \"dnsmasq-dns-864d5fc68c-cws22\" (UID: \"ee60cfe8-e5d9-4e9a-9299-38a14710fcdd\") " pod="openstack/dnsmasq-dns-864d5fc68c-cws22" Nov 29 04:59:12 crc kubenswrapper[4799]: I1129 04:59:12.047537 4799 generic.go:334] "Generic (PLEG): container finished" podID="c0b29191-5d29-44fa-9cde-c2604b5426bb" containerID="e02ae0768cfb16467fe1cee900bb77a9e6b8885860e9ce2e46625c69d51479b2" exitCode=0 Nov 29 04:59:12 crc kubenswrapper[4799]: I1129 04:59:12.047575 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b856c5697-jgvwh" event={"ID":"c0b29191-5d29-44fa-9cde-c2604b5426bb","Type":"ContainerDied","Data":"e02ae0768cfb16467fe1cee900bb77a9e6b8885860e9ce2e46625c69d51479b2"} Nov 29 04:59:12 crc kubenswrapper[4799]: I1129 04:59:12.070894 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee60cfe8-e5d9-4e9a-9299-38a14710fcdd-dns-svc\") pod \"dnsmasq-dns-864d5fc68c-cws22\" (UID: \"ee60cfe8-e5d9-4e9a-9299-38a14710fcdd\") " pod="openstack/dnsmasq-dns-864d5fc68c-cws22" Nov 29 04:59:12 crc kubenswrapper[4799]: I1129 04:59:12.070967 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ee60cfe8-e5d9-4e9a-9299-38a14710fcdd-ovsdbserver-sb\") pod \"dnsmasq-dns-864d5fc68c-cws22\" (UID: \"ee60cfe8-e5d9-4e9a-9299-38a14710fcdd\") " pod="openstack/dnsmasq-dns-864d5fc68c-cws22" Nov 29 04:59:12 crc kubenswrapper[4799]: I1129 04:59:12.070996 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87595\" (UniqueName: \"kubernetes.io/projected/ee60cfe8-e5d9-4e9a-9299-38a14710fcdd-kube-api-access-87595\") pod \"dnsmasq-dns-864d5fc68c-cws22\" (UID: \"ee60cfe8-e5d9-4e9a-9299-38a14710fcdd\") " pod="openstack/dnsmasq-dns-864d5fc68c-cws22" Nov 29 04:59:12 crc kubenswrapper[4799]: I1129 04:59:12.071020 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ee60cfe8-e5d9-4e9a-9299-38a14710fcdd-ovsdbserver-nb\") pod \"dnsmasq-dns-864d5fc68c-cws22\" (UID: \"ee60cfe8-e5d9-4e9a-9299-38a14710fcdd\") " pod="openstack/dnsmasq-dns-864d5fc68c-cws22" Nov 29 04:59:12 crc kubenswrapper[4799]: I1129 04:59:12.071039 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ee60cfe8-e5d9-4e9a-9299-38a14710fcdd-openstack-edpm-ipam\") pod \"dnsmasq-dns-864d5fc68c-cws22\" (UID: \"ee60cfe8-e5d9-4e9a-9299-38a14710fcdd\") " pod="openstack/dnsmasq-dns-864d5fc68c-cws22" Nov 29 04:59:12 crc kubenswrapper[4799]: I1129 04:59:12.071090 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee60cfe8-e5d9-4e9a-9299-38a14710fcdd-config\") pod \"dnsmasq-dns-864d5fc68c-cws22\" (UID: \"ee60cfe8-e5d9-4e9a-9299-38a14710fcdd\") " pod="openstack/dnsmasq-dns-864d5fc68c-cws22" Nov 29 04:59:12 crc kubenswrapper[4799]: I1129 04:59:12.073600 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee60cfe8-e5d9-4e9a-9299-38a14710fcdd-config\") pod \"dnsmasq-dns-864d5fc68c-cws22\" (UID: \"ee60cfe8-e5d9-4e9a-9299-38a14710fcdd\") " pod="openstack/dnsmasq-dns-864d5fc68c-cws22" Nov 29 04:59:12 crc kubenswrapper[4799]: I1129 04:59:12.073655 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ee60cfe8-e5d9-4e9a-9299-38a14710fcdd-ovsdbserver-nb\") pod \"dnsmasq-dns-864d5fc68c-cws22\" (UID: \"ee60cfe8-e5d9-4e9a-9299-38a14710fcdd\") " pod="openstack/dnsmasq-dns-864d5fc68c-cws22" Nov 29 04:59:12 crc kubenswrapper[4799]: I1129 04:59:12.074264 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ee60cfe8-e5d9-4e9a-9299-38a14710fcdd-ovsdbserver-sb\") pod \"dnsmasq-dns-864d5fc68c-cws22\" (UID: \"ee60cfe8-e5d9-4e9a-9299-38a14710fcdd\") " pod="openstack/dnsmasq-dns-864d5fc68c-cws22" Nov 29 04:59:12 crc kubenswrapper[4799]: I1129 04:59:12.074988 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee60cfe8-e5d9-4e9a-9299-38a14710fcdd-dns-svc\") pod \"dnsmasq-dns-864d5fc68c-cws22\" (UID: \"ee60cfe8-e5d9-4e9a-9299-38a14710fcdd\") " pod="openstack/dnsmasq-dns-864d5fc68c-cws22" Nov 29 04:59:12 crc kubenswrapper[4799]: I1129 04:59:12.075381 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ee60cfe8-e5d9-4e9a-9299-38a14710fcdd-openstack-edpm-ipam\") pod \"dnsmasq-dns-864d5fc68c-cws22\" (UID: \"ee60cfe8-e5d9-4e9a-9299-38a14710fcdd\") " pod="openstack/dnsmasq-dns-864d5fc68c-cws22" Nov 29 04:59:12 crc kubenswrapper[4799]: I1129 04:59:12.095008 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87595\" (UniqueName: \"kubernetes.io/projected/ee60cfe8-e5d9-4e9a-9299-38a14710fcdd-kube-api-access-87595\") pod \"dnsmasq-dns-864d5fc68c-cws22\" (UID: \"ee60cfe8-e5d9-4e9a-9299-38a14710fcdd\") " pod="openstack/dnsmasq-dns-864d5fc68c-cws22" Nov 29 04:59:12 crc kubenswrapper[4799]: I1129 04:59:12.145194 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b856c5697-jgvwh" Nov 29 04:59:12 crc kubenswrapper[4799]: I1129 04:59:12.152219 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-864d5fc68c-cws22" Nov 29 04:59:12 crc kubenswrapper[4799]: I1129 04:59:12.274712 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0b29191-5d29-44fa-9cde-c2604b5426bb-config\") pod \"c0b29191-5d29-44fa-9cde-c2604b5426bb\" (UID: \"c0b29191-5d29-44fa-9cde-c2604b5426bb\") " Nov 29 04:59:12 crc kubenswrapper[4799]: I1129 04:59:12.274841 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c0b29191-5d29-44fa-9cde-c2604b5426bb-ovsdbserver-sb\") pod \"c0b29191-5d29-44fa-9cde-c2604b5426bb\" (UID: \"c0b29191-5d29-44fa-9cde-c2604b5426bb\") " Nov 29 04:59:12 crc kubenswrapper[4799]: I1129 04:59:12.274934 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c0b29191-5d29-44fa-9cde-c2604b5426bb-dns-svc\") pod \"c0b29191-5d29-44fa-9cde-c2604b5426bb\" (UID: \"c0b29191-5d29-44fa-9cde-c2604b5426bb\") " Nov 29 04:59:12 crc kubenswrapper[4799]: I1129 04:59:12.274971 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hl42r\" (UniqueName: \"kubernetes.io/projected/c0b29191-5d29-44fa-9cde-c2604b5426bb-kube-api-access-hl42r\") pod \"c0b29191-5d29-44fa-9cde-c2604b5426bb\" (UID: \"c0b29191-5d29-44fa-9cde-c2604b5426bb\") " Nov 29 04:59:12 crc kubenswrapper[4799]: I1129 04:59:12.275010 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c0b29191-5d29-44fa-9cde-c2604b5426bb-ovsdbserver-nb\") pod \"c0b29191-5d29-44fa-9cde-c2604b5426bb\" (UID: \"c0b29191-5d29-44fa-9cde-c2604b5426bb\") " Nov 29 04:59:12 crc kubenswrapper[4799]: I1129 04:59:12.280390 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0b29191-5d29-44fa-9cde-c2604b5426bb-kube-api-access-hl42r" (OuterVolumeSpecName: "kube-api-access-hl42r") pod "c0b29191-5d29-44fa-9cde-c2604b5426bb" (UID: "c0b29191-5d29-44fa-9cde-c2604b5426bb"). InnerVolumeSpecName "kube-api-access-hl42r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:59:12 crc kubenswrapper[4799]: I1129 04:59:12.325421 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c0b29191-5d29-44fa-9cde-c2604b5426bb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c0b29191-5d29-44fa-9cde-c2604b5426bb" (UID: "c0b29191-5d29-44fa-9cde-c2604b5426bb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:59:12 crc kubenswrapper[4799]: I1129 04:59:12.337527 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c0b29191-5d29-44fa-9cde-c2604b5426bb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c0b29191-5d29-44fa-9cde-c2604b5426bb" (UID: "c0b29191-5d29-44fa-9cde-c2604b5426bb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:59:12 crc kubenswrapper[4799]: I1129 04:59:12.344263 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c0b29191-5d29-44fa-9cde-c2604b5426bb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c0b29191-5d29-44fa-9cde-c2604b5426bb" (UID: "c0b29191-5d29-44fa-9cde-c2604b5426bb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:59:12 crc kubenswrapper[4799]: I1129 04:59:12.349653 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c0b29191-5d29-44fa-9cde-c2604b5426bb-config" (OuterVolumeSpecName: "config") pod "c0b29191-5d29-44fa-9cde-c2604b5426bb" (UID: "c0b29191-5d29-44fa-9cde-c2604b5426bb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:59:12 crc kubenswrapper[4799]: I1129 04:59:12.377103 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0b29191-5d29-44fa-9cde-c2604b5426bb-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:59:12 crc kubenswrapper[4799]: I1129 04:59:12.377140 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c0b29191-5d29-44fa-9cde-c2604b5426bb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 29 04:59:12 crc kubenswrapper[4799]: I1129 04:59:12.377153 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c0b29191-5d29-44fa-9cde-c2604b5426bb-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 29 04:59:12 crc kubenswrapper[4799]: I1129 04:59:12.377162 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hl42r\" (UniqueName: \"kubernetes.io/projected/c0b29191-5d29-44fa-9cde-c2604b5426bb-kube-api-access-hl42r\") on node \"crc\" DevicePath \"\"" Nov 29 04:59:12 crc kubenswrapper[4799]: I1129 04:59:12.377174 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c0b29191-5d29-44fa-9cde-c2604b5426bb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 29 04:59:12 crc kubenswrapper[4799]: I1129 04:59:12.674098 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-864d5fc68c-cws22"] Nov 29 04:59:13 crc kubenswrapper[4799]: I1129 04:59:13.059557 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b856c5697-jgvwh" event={"ID":"c0b29191-5d29-44fa-9cde-c2604b5426bb","Type":"ContainerDied","Data":"15e83276f8610bc12c801e4748f4a50e278929c7450ae3f6c9c44af7e9a9fbf9"} Nov 29 04:59:13 crc kubenswrapper[4799]: I1129 04:59:13.060004 4799 scope.go:117] "RemoveContainer" containerID="e02ae0768cfb16467fe1cee900bb77a9e6b8885860e9ce2e46625c69d51479b2" Nov 29 04:59:13 crc kubenswrapper[4799]: I1129 04:59:13.059592 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b856c5697-jgvwh" Nov 29 04:59:13 crc kubenswrapper[4799]: I1129 04:59:13.062216 4799 generic.go:334] "Generic (PLEG): container finished" podID="ee60cfe8-e5d9-4e9a-9299-38a14710fcdd" containerID="f766d10c1e89373de58494f3c5230999c5644629c91ff54a32ee511452b165c5" exitCode=0 Nov 29 04:59:13 crc kubenswrapper[4799]: I1129 04:59:13.062241 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-864d5fc68c-cws22" event={"ID":"ee60cfe8-e5d9-4e9a-9299-38a14710fcdd","Type":"ContainerDied","Data":"f766d10c1e89373de58494f3c5230999c5644629c91ff54a32ee511452b165c5"} Nov 29 04:59:13 crc kubenswrapper[4799]: I1129 04:59:13.062255 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-864d5fc68c-cws22" event={"ID":"ee60cfe8-e5d9-4e9a-9299-38a14710fcdd","Type":"ContainerStarted","Data":"8938a1302f500518ad2f3852e6fdc2286b8b29860d55d2dc2bc6c920623bdf5d"} Nov 29 04:59:13 crc kubenswrapper[4799]: I1129 04:59:13.235753 4799 scope.go:117] "RemoveContainer" containerID="230e52c4037b2f744e5c683c4c87ce6888eeef33464b66a4c1252f5b3054cf7e" Nov 29 04:59:13 crc kubenswrapper[4799]: I1129 04:59:13.273680 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b856c5697-jgvwh"] Nov 29 04:59:13 crc kubenswrapper[4799]: I1129 04:59:13.289928 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b856c5697-jgvwh"] Nov 29 04:59:14 crc kubenswrapper[4799]: I1129 04:59:14.074124 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-864d5fc68c-cws22" event={"ID":"ee60cfe8-e5d9-4e9a-9299-38a14710fcdd","Type":"ContainerStarted","Data":"c51c2754d228077a09b73dce28006df605dc6d539e565145ce5b6edffcfe0de8"} Nov 29 04:59:14 crc kubenswrapper[4799]: I1129 04:59:14.074640 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-864d5fc68c-cws22" Nov 29 04:59:14 crc kubenswrapper[4799]: I1129 04:59:14.103305 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-864d5fc68c-cws22" podStartSLOduration=3.103285015 podStartE2EDuration="3.103285015s" podCreationTimestamp="2025-11-29 04:59:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:59:14.103008858 +0000 UTC m=+1229.745939268" watchObservedRunningTime="2025-11-29 04:59:14.103285015 +0000 UTC m=+1229.746215415" Nov 29 04:59:14 crc kubenswrapper[4799]: I1129 04:59:14.667834 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0b29191-5d29-44fa-9cde-c2604b5426bb" path="/var/lib/kubelet/pods/c0b29191-5d29-44fa-9cde-c2604b5426bb/volumes" Nov 29 04:59:17 crc kubenswrapper[4799]: I1129 04:59:17.361564 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mzjpx"] Nov 29 04:59:17 crc kubenswrapper[4799]: E1129 04:59:17.362035 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0b29191-5d29-44fa-9cde-c2604b5426bb" containerName="init" Nov 29 04:59:17 crc kubenswrapper[4799]: I1129 04:59:17.362050 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0b29191-5d29-44fa-9cde-c2604b5426bb" containerName="init" Nov 29 04:59:17 crc kubenswrapper[4799]: E1129 04:59:17.362066 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0b29191-5d29-44fa-9cde-c2604b5426bb" containerName="dnsmasq-dns" Nov 29 04:59:17 crc kubenswrapper[4799]: I1129 04:59:17.362074 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0b29191-5d29-44fa-9cde-c2604b5426bb" containerName="dnsmasq-dns" Nov 29 04:59:17 crc kubenswrapper[4799]: I1129 04:59:17.362298 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0b29191-5d29-44fa-9cde-c2604b5426bb" containerName="dnsmasq-dns" Nov 29 04:59:17 crc kubenswrapper[4799]: I1129 04:59:17.363076 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mzjpx" Nov 29 04:59:17 crc kubenswrapper[4799]: I1129 04:59:17.365176 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 29 04:59:17 crc kubenswrapper[4799]: I1129 04:59:17.365283 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 29 04:59:17 crc kubenswrapper[4799]: I1129 04:59:17.365483 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 29 04:59:17 crc kubenswrapper[4799]: I1129 04:59:17.369082 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ljwtk" Nov 29 04:59:17 crc kubenswrapper[4799]: I1129 04:59:17.380514 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mzjpx"] Nov 29 04:59:17 crc kubenswrapper[4799]: I1129 04:59:17.471660 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa56b0dc-b80d-44d7-8c89-0081d33bba8f-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-mzjpx\" (UID: \"aa56b0dc-b80d-44d7-8c89-0081d33bba8f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mzjpx" Nov 29 04:59:17 crc kubenswrapper[4799]: I1129 04:59:17.471775 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa56b0dc-b80d-44d7-8c89-0081d33bba8f-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-mzjpx\" (UID: \"aa56b0dc-b80d-44d7-8c89-0081d33bba8f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mzjpx" Nov 29 04:59:17 crc kubenswrapper[4799]: I1129 04:59:17.471932 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa56b0dc-b80d-44d7-8c89-0081d33bba8f-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-mzjpx\" (UID: \"aa56b0dc-b80d-44d7-8c89-0081d33bba8f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mzjpx" Nov 29 04:59:17 crc kubenswrapper[4799]: I1129 04:59:17.471975 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkh26\" (UniqueName: \"kubernetes.io/projected/aa56b0dc-b80d-44d7-8c89-0081d33bba8f-kube-api-access-dkh26\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-mzjpx\" (UID: \"aa56b0dc-b80d-44d7-8c89-0081d33bba8f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mzjpx" Nov 29 04:59:17 crc kubenswrapper[4799]: I1129 04:59:17.573436 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkh26\" (UniqueName: \"kubernetes.io/projected/aa56b0dc-b80d-44d7-8c89-0081d33bba8f-kube-api-access-dkh26\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-mzjpx\" (UID: \"aa56b0dc-b80d-44d7-8c89-0081d33bba8f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mzjpx" Nov 29 04:59:17 crc kubenswrapper[4799]: I1129 04:59:17.573521 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa56b0dc-b80d-44d7-8c89-0081d33bba8f-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-mzjpx\" (UID: \"aa56b0dc-b80d-44d7-8c89-0081d33bba8f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mzjpx" Nov 29 04:59:17 crc kubenswrapper[4799]: I1129 04:59:17.573601 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa56b0dc-b80d-44d7-8c89-0081d33bba8f-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-mzjpx\" (UID: \"aa56b0dc-b80d-44d7-8c89-0081d33bba8f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mzjpx" Nov 29 04:59:17 crc kubenswrapper[4799]: I1129 04:59:17.573687 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa56b0dc-b80d-44d7-8c89-0081d33bba8f-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-mzjpx\" (UID: \"aa56b0dc-b80d-44d7-8c89-0081d33bba8f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mzjpx" Nov 29 04:59:17 crc kubenswrapper[4799]: I1129 04:59:17.586688 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa56b0dc-b80d-44d7-8c89-0081d33bba8f-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-mzjpx\" (UID: \"aa56b0dc-b80d-44d7-8c89-0081d33bba8f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mzjpx" Nov 29 04:59:17 crc kubenswrapper[4799]: I1129 04:59:17.594017 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa56b0dc-b80d-44d7-8c89-0081d33bba8f-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-mzjpx\" (UID: \"aa56b0dc-b80d-44d7-8c89-0081d33bba8f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mzjpx" Nov 29 04:59:17 crc kubenswrapper[4799]: I1129 04:59:17.594713 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa56b0dc-b80d-44d7-8c89-0081d33bba8f-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-mzjpx\" (UID: \"aa56b0dc-b80d-44d7-8c89-0081d33bba8f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mzjpx" Nov 29 04:59:17 crc kubenswrapper[4799]: I1129 04:59:17.605904 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkh26\" (UniqueName: \"kubernetes.io/projected/aa56b0dc-b80d-44d7-8c89-0081d33bba8f-kube-api-access-dkh26\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-mzjpx\" (UID: \"aa56b0dc-b80d-44d7-8c89-0081d33bba8f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mzjpx" Nov 29 04:59:17 crc kubenswrapper[4799]: I1129 04:59:17.680949 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mzjpx" Nov 29 04:59:18 crc kubenswrapper[4799]: I1129 04:59:18.295570 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mzjpx"] Nov 29 04:59:18 crc kubenswrapper[4799]: I1129 04:59:18.298104 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 29 04:59:19 crc kubenswrapper[4799]: I1129 04:59:19.120317 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mzjpx" event={"ID":"aa56b0dc-b80d-44d7-8c89-0081d33bba8f","Type":"ContainerStarted","Data":"748a6a902ea6bf08a00ec3f6c32308f7b8af9db799d98a259d31286b042df1f5"} Nov 29 04:59:22 crc kubenswrapper[4799]: I1129 04:59:22.154069 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-864d5fc68c-cws22" Nov 29 04:59:22 crc kubenswrapper[4799]: I1129 04:59:22.230860 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6447ccbd8f-cvd9n"] Nov 29 04:59:22 crc kubenswrapper[4799]: I1129 04:59:22.231238 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6447ccbd8f-cvd9n" podUID="0a332883-d0b4-483f-8ae9-93cf97d2815a" containerName="dnsmasq-dns" containerID="cri-o://eadd10961388f58d49c48365e68b1dbc5b6715e2bf945f0943d0cef0351a0261" gracePeriod=10 Nov 29 04:59:23 crc kubenswrapper[4799]: I1129 04:59:23.175118 4799 generic.go:334] "Generic (PLEG): container finished" podID="0a332883-d0b4-483f-8ae9-93cf97d2815a" containerID="eadd10961388f58d49c48365e68b1dbc5b6715e2bf945f0943d0cef0351a0261" exitCode=0 Nov 29 04:59:23 crc kubenswrapper[4799]: I1129 04:59:23.175174 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6447ccbd8f-cvd9n" event={"ID":"0a332883-d0b4-483f-8ae9-93cf97d2815a","Type":"ContainerDied","Data":"eadd10961388f58d49c48365e68b1dbc5b6715e2bf945f0943d0cef0351a0261"} Nov 29 04:59:26 crc kubenswrapper[4799]: I1129 04:59:26.498260 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6447ccbd8f-cvd9n" podUID="0a332883-d0b4-483f-8ae9-93cf97d2815a" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.188:5353: connect: connection refused" Nov 29 04:59:28 crc kubenswrapper[4799]: I1129 04:59:28.114036 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6447ccbd8f-cvd9n" Nov 29 04:59:28 crc kubenswrapper[4799]: I1129 04:59:28.208007 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0a332883-d0b4-483f-8ae9-93cf97d2815a-ovsdbserver-sb\") pod \"0a332883-d0b4-483f-8ae9-93cf97d2815a\" (UID: \"0a332883-d0b4-483f-8ae9-93cf97d2815a\") " Nov 29 04:59:28 crc kubenswrapper[4799]: I1129 04:59:28.208078 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a332883-d0b4-483f-8ae9-93cf97d2815a-dns-svc\") pod \"0a332883-d0b4-483f-8ae9-93cf97d2815a\" (UID: \"0a332883-d0b4-483f-8ae9-93cf97d2815a\") " Nov 29 04:59:28 crc kubenswrapper[4799]: I1129 04:59:28.208154 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a332883-d0b4-483f-8ae9-93cf97d2815a-config\") pod \"0a332883-d0b4-483f-8ae9-93cf97d2815a\" (UID: \"0a332883-d0b4-483f-8ae9-93cf97d2815a\") " Nov 29 04:59:28 crc kubenswrapper[4799]: I1129 04:59:28.208221 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k5djw\" (UniqueName: \"kubernetes.io/projected/0a332883-d0b4-483f-8ae9-93cf97d2815a-kube-api-access-k5djw\") pod \"0a332883-d0b4-483f-8ae9-93cf97d2815a\" (UID: \"0a332883-d0b4-483f-8ae9-93cf97d2815a\") " Nov 29 04:59:28 crc kubenswrapper[4799]: I1129 04:59:28.208240 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0a332883-d0b4-483f-8ae9-93cf97d2815a-ovsdbserver-nb\") pod \"0a332883-d0b4-483f-8ae9-93cf97d2815a\" (UID: \"0a332883-d0b4-483f-8ae9-93cf97d2815a\") " Nov 29 04:59:28 crc kubenswrapper[4799]: I1129 04:59:28.208271 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0a332883-d0b4-483f-8ae9-93cf97d2815a-openstack-edpm-ipam\") pod \"0a332883-d0b4-483f-8ae9-93cf97d2815a\" (UID: \"0a332883-d0b4-483f-8ae9-93cf97d2815a\") " Nov 29 04:59:28 crc kubenswrapper[4799]: I1129 04:59:28.212945 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a332883-d0b4-483f-8ae9-93cf97d2815a-kube-api-access-k5djw" (OuterVolumeSpecName: "kube-api-access-k5djw") pod "0a332883-d0b4-483f-8ae9-93cf97d2815a" (UID: "0a332883-d0b4-483f-8ae9-93cf97d2815a"). InnerVolumeSpecName "kube-api-access-k5djw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:59:28 crc kubenswrapper[4799]: I1129 04:59:28.230894 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6447ccbd8f-cvd9n" event={"ID":"0a332883-d0b4-483f-8ae9-93cf97d2815a","Type":"ContainerDied","Data":"1612a74094cc52052e6752ca2cecdadaf14119c22173fa9f33e3a99833e28b8f"} Nov 29 04:59:28 crc kubenswrapper[4799]: I1129 04:59:28.230964 4799 scope.go:117] "RemoveContainer" containerID="eadd10961388f58d49c48365e68b1dbc5b6715e2bf945f0943d0cef0351a0261" Nov 29 04:59:28 crc kubenswrapper[4799]: I1129 04:59:28.230963 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6447ccbd8f-cvd9n" Nov 29 04:59:28 crc kubenswrapper[4799]: I1129 04:59:28.257060 4799 scope.go:117] "RemoveContainer" containerID="7d1a52c97518110d33c0ca415378f6e995179958cd21b2f67174b585b8097df2" Nov 29 04:59:28 crc kubenswrapper[4799]: I1129 04:59:28.259948 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a332883-d0b4-483f-8ae9-93cf97d2815a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0a332883-d0b4-483f-8ae9-93cf97d2815a" (UID: "0a332883-d0b4-483f-8ae9-93cf97d2815a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:59:28 crc kubenswrapper[4799]: I1129 04:59:28.260683 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a332883-d0b4-483f-8ae9-93cf97d2815a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0a332883-d0b4-483f-8ae9-93cf97d2815a" (UID: "0a332883-d0b4-483f-8ae9-93cf97d2815a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:59:28 crc kubenswrapper[4799]: I1129 04:59:28.261357 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a332883-d0b4-483f-8ae9-93cf97d2815a-config" (OuterVolumeSpecName: "config") pod "0a332883-d0b4-483f-8ae9-93cf97d2815a" (UID: "0a332883-d0b4-483f-8ae9-93cf97d2815a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:59:28 crc kubenswrapper[4799]: I1129 04:59:28.267630 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a332883-d0b4-483f-8ae9-93cf97d2815a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0a332883-d0b4-483f-8ae9-93cf97d2815a" (UID: "0a332883-d0b4-483f-8ae9-93cf97d2815a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:59:28 crc kubenswrapper[4799]: I1129 04:59:28.280519 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a332883-d0b4-483f-8ae9-93cf97d2815a-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "0a332883-d0b4-483f-8ae9-93cf97d2815a" (UID: "0a332883-d0b4-483f-8ae9-93cf97d2815a"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 04:59:28 crc kubenswrapper[4799]: I1129 04:59:28.309814 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a332883-d0b4-483f-8ae9-93cf97d2815a-config\") on node \"crc\" DevicePath \"\"" Nov 29 04:59:28 crc kubenswrapper[4799]: I1129 04:59:28.309842 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k5djw\" (UniqueName: \"kubernetes.io/projected/0a332883-d0b4-483f-8ae9-93cf97d2815a-kube-api-access-k5djw\") on node \"crc\" DevicePath \"\"" Nov 29 04:59:28 crc kubenswrapper[4799]: I1129 04:59:28.309853 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0a332883-d0b4-483f-8ae9-93cf97d2815a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 29 04:59:28 crc kubenswrapper[4799]: I1129 04:59:28.309861 4799 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0a332883-d0b4-483f-8ae9-93cf97d2815a-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 29 04:59:28 crc kubenswrapper[4799]: I1129 04:59:28.309871 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0a332883-d0b4-483f-8ae9-93cf97d2815a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 29 04:59:28 crc kubenswrapper[4799]: I1129 04:59:28.309879 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a332883-d0b4-483f-8ae9-93cf97d2815a-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 29 04:59:28 crc kubenswrapper[4799]: I1129 04:59:28.581285 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6447ccbd8f-cvd9n"] Nov 29 04:59:28 crc kubenswrapper[4799]: I1129 04:59:28.594572 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6447ccbd8f-cvd9n"] Nov 29 04:59:28 crc kubenswrapper[4799]: I1129 04:59:28.676173 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a332883-d0b4-483f-8ae9-93cf97d2815a" path="/var/lib/kubelet/pods/0a332883-d0b4-483f-8ae9-93cf97d2815a/volumes" Nov 29 04:59:29 crc kubenswrapper[4799]: I1129 04:59:29.240762 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mzjpx" event={"ID":"aa56b0dc-b80d-44d7-8c89-0081d33bba8f","Type":"ContainerStarted","Data":"3d6883d8d8a872a0fc6be069ceab4743b12e54414dcbe264bf35dda2e1182b7e"} Nov 29 04:59:30 crc kubenswrapper[4799]: I1129 04:59:30.267518 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mzjpx" podStartSLOduration=3.671180624 podStartE2EDuration="13.267498504s" podCreationTimestamp="2025-11-29 04:59:17 +0000 UTC" firstStartedPulling="2025-11-29 04:59:18.297729021 +0000 UTC m=+1233.940659421" lastFinishedPulling="2025-11-29 04:59:27.894046901 +0000 UTC m=+1243.536977301" observedRunningTime="2025-11-29 04:59:30.262560693 +0000 UTC m=+1245.905491093" watchObservedRunningTime="2025-11-29 04:59:30.267498504 +0000 UTC m=+1245.910428904" Nov 29 04:59:35 crc kubenswrapper[4799]: I1129 04:59:35.303884 4799 generic.go:334] "Generic (PLEG): container finished" podID="0bfecf67-24fe-42fa-bad5-4a597ad9d3af" containerID="c3f70cf40e8091597499a0c128f7230a1c8d1dca9ef99085ae1033299fe6d684" exitCode=0 Nov 29 04:59:35 crc kubenswrapper[4799]: I1129 04:59:35.303954 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0bfecf67-24fe-42fa-bad5-4a597ad9d3af","Type":"ContainerDied","Data":"c3f70cf40e8091597499a0c128f7230a1c8d1dca9ef99085ae1033299fe6d684"} Nov 29 04:59:35 crc kubenswrapper[4799]: I1129 04:59:35.307400 4799 generic.go:334] "Generic (PLEG): container finished" podID="1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac" containerID="abe77cc4d36024c6b591d4d16586151c95686b7404c44a477d7d08d78de919ee" exitCode=0 Nov 29 04:59:35 crc kubenswrapper[4799]: I1129 04:59:35.307470 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac","Type":"ContainerDied","Data":"abe77cc4d36024c6b591d4d16586151c95686b7404c44a477d7d08d78de919ee"} Nov 29 04:59:37 crc kubenswrapper[4799]: I1129 04:59:37.329141 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0bfecf67-24fe-42fa-bad5-4a597ad9d3af","Type":"ContainerStarted","Data":"26c819fbf8a7b1959adbdde2fd18095d21091880848df4f21fa072389a875984"} Nov 29 04:59:37 crc kubenswrapper[4799]: I1129 04:59:37.333088 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac","Type":"ContainerStarted","Data":"16746ee48b332829fc94abe3d3c0c4d5988f0f4d560461289e4f501454ac3e5c"} Nov 29 04:59:38 crc kubenswrapper[4799]: I1129 04:59:38.342577 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 29 04:59:38 crc kubenswrapper[4799]: I1129 04:59:38.342933 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 29 04:59:38 crc kubenswrapper[4799]: I1129 04:59:38.364862 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=40.364847274 podStartE2EDuration="40.364847274s" podCreationTimestamp="2025-11-29 04:58:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:59:38.364359372 +0000 UTC m=+1254.007289852" watchObservedRunningTime="2025-11-29 04:59:38.364847274 +0000 UTC m=+1254.007777674" Nov 29 04:59:38 crc kubenswrapper[4799]: I1129 04:59:38.395297 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=40.395257779 podStartE2EDuration="40.395257779s" podCreationTimestamp="2025-11-29 04:58:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 04:59:38.384691179 +0000 UTC m=+1254.027621589" watchObservedRunningTime="2025-11-29 04:59:38.395257779 +0000 UTC m=+1254.038188179" Nov 29 04:59:42 crc kubenswrapper[4799]: I1129 04:59:42.375617 4799 generic.go:334] "Generic (PLEG): container finished" podID="aa56b0dc-b80d-44d7-8c89-0081d33bba8f" containerID="3d6883d8d8a872a0fc6be069ceab4743b12e54414dcbe264bf35dda2e1182b7e" exitCode=0 Nov 29 04:59:42 crc kubenswrapper[4799]: I1129 04:59:42.375701 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mzjpx" event={"ID":"aa56b0dc-b80d-44d7-8c89-0081d33bba8f","Type":"ContainerDied","Data":"3d6883d8d8a872a0fc6be069ceab4743b12e54414dcbe264bf35dda2e1182b7e"} Nov 29 04:59:43 crc kubenswrapper[4799]: I1129 04:59:43.847327 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mzjpx" Nov 29 04:59:43 crc kubenswrapper[4799]: I1129 04:59:43.939101 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dkh26\" (UniqueName: \"kubernetes.io/projected/aa56b0dc-b80d-44d7-8c89-0081d33bba8f-kube-api-access-dkh26\") pod \"aa56b0dc-b80d-44d7-8c89-0081d33bba8f\" (UID: \"aa56b0dc-b80d-44d7-8c89-0081d33bba8f\") " Nov 29 04:59:43 crc kubenswrapper[4799]: I1129 04:59:43.939181 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa56b0dc-b80d-44d7-8c89-0081d33bba8f-repo-setup-combined-ca-bundle\") pod \"aa56b0dc-b80d-44d7-8c89-0081d33bba8f\" (UID: \"aa56b0dc-b80d-44d7-8c89-0081d33bba8f\") " Nov 29 04:59:43 crc kubenswrapper[4799]: I1129 04:59:43.939210 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa56b0dc-b80d-44d7-8c89-0081d33bba8f-ssh-key\") pod \"aa56b0dc-b80d-44d7-8c89-0081d33bba8f\" (UID: \"aa56b0dc-b80d-44d7-8c89-0081d33bba8f\") " Nov 29 04:59:43 crc kubenswrapper[4799]: I1129 04:59:43.944853 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa56b0dc-b80d-44d7-8c89-0081d33bba8f-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "aa56b0dc-b80d-44d7-8c89-0081d33bba8f" (UID: "aa56b0dc-b80d-44d7-8c89-0081d33bba8f"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:59:43 crc kubenswrapper[4799]: I1129 04:59:43.944924 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa56b0dc-b80d-44d7-8c89-0081d33bba8f-kube-api-access-dkh26" (OuterVolumeSpecName: "kube-api-access-dkh26") pod "aa56b0dc-b80d-44d7-8c89-0081d33bba8f" (UID: "aa56b0dc-b80d-44d7-8c89-0081d33bba8f"). InnerVolumeSpecName "kube-api-access-dkh26". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 04:59:43 crc kubenswrapper[4799]: I1129 04:59:43.970389 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa56b0dc-b80d-44d7-8c89-0081d33bba8f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "aa56b0dc-b80d-44d7-8c89-0081d33bba8f" (UID: "aa56b0dc-b80d-44d7-8c89-0081d33bba8f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:59:44 crc kubenswrapper[4799]: I1129 04:59:44.041297 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa56b0dc-b80d-44d7-8c89-0081d33bba8f-inventory\") pod \"aa56b0dc-b80d-44d7-8c89-0081d33bba8f\" (UID: \"aa56b0dc-b80d-44d7-8c89-0081d33bba8f\") " Nov 29 04:59:44 crc kubenswrapper[4799]: I1129 04:59:44.041945 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dkh26\" (UniqueName: \"kubernetes.io/projected/aa56b0dc-b80d-44d7-8c89-0081d33bba8f-kube-api-access-dkh26\") on node \"crc\" DevicePath \"\"" Nov 29 04:59:44 crc kubenswrapper[4799]: I1129 04:59:44.041973 4799 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa56b0dc-b80d-44d7-8c89-0081d33bba8f-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 04:59:44 crc kubenswrapper[4799]: I1129 04:59:44.041988 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa56b0dc-b80d-44d7-8c89-0081d33bba8f-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 29 04:59:44 crc kubenswrapper[4799]: I1129 04:59:44.081369 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa56b0dc-b80d-44d7-8c89-0081d33bba8f-inventory" (OuterVolumeSpecName: "inventory") pod "aa56b0dc-b80d-44d7-8c89-0081d33bba8f" (UID: "aa56b0dc-b80d-44d7-8c89-0081d33bba8f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 04:59:44 crc kubenswrapper[4799]: I1129 04:59:44.146901 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa56b0dc-b80d-44d7-8c89-0081d33bba8f-inventory\") on node \"crc\" DevicePath \"\"" Nov 29 04:59:44 crc kubenswrapper[4799]: I1129 04:59:44.399023 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mzjpx" event={"ID":"aa56b0dc-b80d-44d7-8c89-0081d33bba8f","Type":"ContainerDied","Data":"748a6a902ea6bf08a00ec3f6c32308f7b8af9db799d98a259d31286b042df1f5"} Nov 29 04:59:44 crc kubenswrapper[4799]: I1129 04:59:44.399076 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="748a6a902ea6bf08a00ec3f6c32308f7b8af9db799d98a259d31286b042df1f5" Nov 29 04:59:44 crc kubenswrapper[4799]: I1129 04:59:44.399083 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mzjpx" Nov 29 04:59:44 crc kubenswrapper[4799]: I1129 04:59:44.531289 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pvn8g"] Nov 29 04:59:44 crc kubenswrapper[4799]: E1129 04:59:44.531673 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a332883-d0b4-483f-8ae9-93cf97d2815a" containerName="init" Nov 29 04:59:44 crc kubenswrapper[4799]: I1129 04:59:44.531692 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a332883-d0b4-483f-8ae9-93cf97d2815a" containerName="init" Nov 29 04:59:44 crc kubenswrapper[4799]: E1129 04:59:44.531716 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a332883-d0b4-483f-8ae9-93cf97d2815a" containerName="dnsmasq-dns" Nov 29 04:59:44 crc kubenswrapper[4799]: I1129 04:59:44.531725 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a332883-d0b4-483f-8ae9-93cf97d2815a" containerName="dnsmasq-dns" Nov 29 04:59:44 crc kubenswrapper[4799]: E1129 04:59:44.531735 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa56b0dc-b80d-44d7-8c89-0081d33bba8f" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 29 04:59:44 crc kubenswrapper[4799]: I1129 04:59:44.531743 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa56b0dc-b80d-44d7-8c89-0081d33bba8f" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 29 04:59:44 crc kubenswrapper[4799]: I1129 04:59:44.531958 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a332883-d0b4-483f-8ae9-93cf97d2815a" containerName="dnsmasq-dns" Nov 29 04:59:44 crc kubenswrapper[4799]: I1129 04:59:44.531982 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa56b0dc-b80d-44d7-8c89-0081d33bba8f" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 29 04:59:44 crc kubenswrapper[4799]: I1129 04:59:44.532526 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pvn8g" Nov 29 04:59:44 crc kubenswrapper[4799]: I1129 04:59:44.535723 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ljwtk" Nov 29 04:59:44 crc kubenswrapper[4799]: I1129 04:59:44.536198 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 29 04:59:44 crc kubenswrapper[4799]: I1129 04:59:44.536447 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 29 04:59:44 crc kubenswrapper[4799]: I1129 04:59:44.537297 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 29 04:59:44 crc kubenswrapper[4799]: I1129 04:59:44.555556 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wr24r\" (UniqueName: \"kubernetes.io/projected/2e328538-a1ef-49dd-9ecd-d83560c36e15-kube-api-access-wr24r\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-pvn8g\" (UID: \"2e328538-a1ef-49dd-9ecd-d83560c36e15\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pvn8g" Nov 29 04:59:44 crc kubenswrapper[4799]: I1129 04:59:44.555917 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2e328538-a1ef-49dd-9ecd-d83560c36e15-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-pvn8g\" (UID: \"2e328538-a1ef-49dd-9ecd-d83560c36e15\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pvn8g" Nov 29 04:59:44 crc kubenswrapper[4799]: I1129 04:59:44.556085 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2e328538-a1ef-49dd-9ecd-d83560c36e15-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-pvn8g\" (UID: \"2e328538-a1ef-49dd-9ecd-d83560c36e15\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pvn8g" Nov 29 04:59:44 crc kubenswrapper[4799]: I1129 04:59:44.556209 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e328538-a1ef-49dd-9ecd-d83560c36e15-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-pvn8g\" (UID: \"2e328538-a1ef-49dd-9ecd-d83560c36e15\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pvn8g" Nov 29 04:59:44 crc kubenswrapper[4799]: I1129 04:59:44.598531 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pvn8g"] Nov 29 04:59:44 crc kubenswrapper[4799]: I1129 04:59:44.657920 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wr24r\" (UniqueName: \"kubernetes.io/projected/2e328538-a1ef-49dd-9ecd-d83560c36e15-kube-api-access-wr24r\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-pvn8g\" (UID: \"2e328538-a1ef-49dd-9ecd-d83560c36e15\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pvn8g" Nov 29 04:59:44 crc kubenswrapper[4799]: I1129 04:59:44.658991 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2e328538-a1ef-49dd-9ecd-d83560c36e15-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-pvn8g\" (UID: \"2e328538-a1ef-49dd-9ecd-d83560c36e15\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pvn8g" Nov 29 04:59:44 crc kubenswrapper[4799]: I1129 04:59:44.659828 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2e328538-a1ef-49dd-9ecd-d83560c36e15-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-pvn8g\" (UID: \"2e328538-a1ef-49dd-9ecd-d83560c36e15\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pvn8g" Nov 29 04:59:44 crc kubenswrapper[4799]: I1129 04:59:44.660156 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e328538-a1ef-49dd-9ecd-d83560c36e15-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-pvn8g\" (UID: \"2e328538-a1ef-49dd-9ecd-d83560c36e15\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pvn8g" Nov 29 04:59:44 crc kubenswrapper[4799]: I1129 04:59:44.669177 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 29 04:59:44 crc kubenswrapper[4799]: I1129 04:59:44.669469 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 29 04:59:44 crc kubenswrapper[4799]: I1129 04:59:44.678378 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e328538-a1ef-49dd-9ecd-d83560c36e15-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-pvn8g\" (UID: \"2e328538-a1ef-49dd-9ecd-d83560c36e15\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pvn8g" Nov 29 04:59:44 crc kubenswrapper[4799]: I1129 04:59:44.687206 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2e328538-a1ef-49dd-9ecd-d83560c36e15-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-pvn8g\" (UID: \"2e328538-a1ef-49dd-9ecd-d83560c36e15\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pvn8g" Nov 29 04:59:44 crc kubenswrapper[4799]: I1129 04:59:44.703290 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2e328538-a1ef-49dd-9ecd-d83560c36e15-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-pvn8g\" (UID: \"2e328538-a1ef-49dd-9ecd-d83560c36e15\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pvn8g" Nov 29 04:59:44 crc kubenswrapper[4799]: I1129 04:59:44.720423 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wr24r\" (UniqueName: \"kubernetes.io/projected/2e328538-a1ef-49dd-9ecd-d83560c36e15-kube-api-access-wr24r\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-pvn8g\" (UID: \"2e328538-a1ef-49dd-9ecd-d83560c36e15\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pvn8g" Nov 29 04:59:44 crc kubenswrapper[4799]: I1129 04:59:44.851618 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ljwtk" Nov 29 04:59:44 crc kubenswrapper[4799]: I1129 04:59:44.859577 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pvn8g" Nov 29 04:59:45 crc kubenswrapper[4799]: I1129 04:59:45.459754 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pvn8g"] Nov 29 04:59:45 crc kubenswrapper[4799]: W1129 04:59:45.467688 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2e328538_a1ef_49dd_9ecd_d83560c36e15.slice/crio-76b5755384e79b20960968a191b64ce8615ea12ab98c2bc8a52c014ade4a78ab WatchSource:0}: Error finding container 76b5755384e79b20960968a191b64ce8615ea12ab98c2bc8a52c014ade4a78ab: Status 404 returned error can't find the container with id 76b5755384e79b20960968a191b64ce8615ea12ab98c2bc8a52c014ade4a78ab Nov 29 04:59:46 crc kubenswrapper[4799]: I1129 04:59:46.179007 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 29 04:59:46 crc kubenswrapper[4799]: I1129 04:59:46.430377 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pvn8g" event={"ID":"2e328538-a1ef-49dd-9ecd-d83560c36e15","Type":"ContainerStarted","Data":"76b5755384e79b20960968a191b64ce8615ea12ab98c2bc8a52c014ade4a78ab"} Nov 29 04:59:47 crc kubenswrapper[4799]: I1129 04:59:47.444074 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pvn8g" event={"ID":"2e328538-a1ef-49dd-9ecd-d83560c36e15","Type":"ContainerStarted","Data":"5d2ced6ff32786bd7820e5889d3ac4340084b3f818960ca1cd8507dbe9a6b4a6"} Nov 29 04:59:47 crc kubenswrapper[4799]: I1129 04:59:47.476249 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pvn8g" podStartSLOduration=2.770582267 podStartE2EDuration="3.476216846s" podCreationTimestamp="2025-11-29 04:59:44 +0000 UTC" firstStartedPulling="2025-11-29 04:59:45.470922383 +0000 UTC m=+1261.113852803" lastFinishedPulling="2025-11-29 04:59:46.176556962 +0000 UTC m=+1261.819487382" observedRunningTime="2025-11-29 04:59:47.466412837 +0000 UTC m=+1263.109343247" watchObservedRunningTime="2025-11-29 04:59:47.476216846 +0000 UTC m=+1263.119147246" Nov 29 04:59:48 crc kubenswrapper[4799]: I1129 04:59:48.521163 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 29 04:59:48 crc kubenswrapper[4799]: I1129 04:59:48.639019 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 29 05:00:00 crc kubenswrapper[4799]: I1129 05:00:00.160072 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406540-xd5kj"] Nov 29 05:00:00 crc kubenswrapper[4799]: I1129 05:00:00.163141 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406540-xd5kj" Nov 29 05:00:00 crc kubenswrapper[4799]: I1129 05:00:00.166415 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 29 05:00:00 crc kubenswrapper[4799]: I1129 05:00:00.168151 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 29 05:00:00 crc kubenswrapper[4799]: I1129 05:00:00.180774 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406540-xd5kj"] Nov 29 05:00:00 crc kubenswrapper[4799]: I1129 05:00:00.328802 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e3ff8087-614c-409e-8238-384a44cdc990-config-volume\") pod \"collect-profiles-29406540-xd5kj\" (UID: \"e3ff8087-614c-409e-8238-384a44cdc990\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406540-xd5kj" Nov 29 05:00:00 crc kubenswrapper[4799]: I1129 05:00:00.329251 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmmkt\" (UniqueName: \"kubernetes.io/projected/e3ff8087-614c-409e-8238-384a44cdc990-kube-api-access-zmmkt\") pod \"collect-profiles-29406540-xd5kj\" (UID: \"e3ff8087-614c-409e-8238-384a44cdc990\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406540-xd5kj" Nov 29 05:00:00 crc kubenswrapper[4799]: I1129 05:00:00.329415 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e3ff8087-614c-409e-8238-384a44cdc990-secret-volume\") pod \"collect-profiles-29406540-xd5kj\" (UID: \"e3ff8087-614c-409e-8238-384a44cdc990\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406540-xd5kj" Nov 29 05:00:00 crc kubenswrapper[4799]: I1129 05:00:00.431564 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmmkt\" (UniqueName: \"kubernetes.io/projected/e3ff8087-614c-409e-8238-384a44cdc990-kube-api-access-zmmkt\") pod \"collect-profiles-29406540-xd5kj\" (UID: \"e3ff8087-614c-409e-8238-384a44cdc990\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406540-xd5kj" Nov 29 05:00:00 crc kubenswrapper[4799]: I1129 05:00:00.431667 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e3ff8087-614c-409e-8238-384a44cdc990-secret-volume\") pod \"collect-profiles-29406540-xd5kj\" (UID: \"e3ff8087-614c-409e-8238-384a44cdc990\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406540-xd5kj" Nov 29 05:00:00 crc kubenswrapper[4799]: I1129 05:00:00.431861 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e3ff8087-614c-409e-8238-384a44cdc990-config-volume\") pod \"collect-profiles-29406540-xd5kj\" (UID: \"e3ff8087-614c-409e-8238-384a44cdc990\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406540-xd5kj" Nov 29 05:00:00 crc kubenswrapper[4799]: I1129 05:00:00.433337 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e3ff8087-614c-409e-8238-384a44cdc990-config-volume\") pod \"collect-profiles-29406540-xd5kj\" (UID: \"e3ff8087-614c-409e-8238-384a44cdc990\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406540-xd5kj" Nov 29 05:00:00 crc kubenswrapper[4799]: I1129 05:00:00.453764 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e3ff8087-614c-409e-8238-384a44cdc990-secret-volume\") pod \"collect-profiles-29406540-xd5kj\" (UID: \"e3ff8087-614c-409e-8238-384a44cdc990\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406540-xd5kj" Nov 29 05:00:00 crc kubenswrapper[4799]: I1129 05:00:00.456068 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmmkt\" (UniqueName: \"kubernetes.io/projected/e3ff8087-614c-409e-8238-384a44cdc990-kube-api-access-zmmkt\") pod \"collect-profiles-29406540-xd5kj\" (UID: \"e3ff8087-614c-409e-8238-384a44cdc990\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406540-xd5kj" Nov 29 05:00:00 crc kubenswrapper[4799]: I1129 05:00:00.506108 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406540-xd5kj" Nov 29 05:00:01 crc kubenswrapper[4799]: I1129 05:00:01.006913 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406540-xd5kj"] Nov 29 05:00:01 crc kubenswrapper[4799]: I1129 05:00:01.617088 4799 generic.go:334] "Generic (PLEG): container finished" podID="e3ff8087-614c-409e-8238-384a44cdc990" containerID="f038d0ded959a6574153aa96512043f347f81bf952618e47406a40b4a8b7221b" exitCode=0 Nov 29 05:00:01 crc kubenswrapper[4799]: I1129 05:00:01.617468 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29406540-xd5kj" event={"ID":"e3ff8087-614c-409e-8238-384a44cdc990","Type":"ContainerDied","Data":"f038d0ded959a6574153aa96512043f347f81bf952618e47406a40b4a8b7221b"} Nov 29 05:00:01 crc kubenswrapper[4799]: I1129 05:00:01.617500 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29406540-xd5kj" event={"ID":"e3ff8087-614c-409e-8238-384a44cdc990","Type":"ContainerStarted","Data":"f3220f9d699cfc22bf09744d5e28fa265bd01aa71a93813052adba6f7160de69"} Nov 29 05:00:02 crc kubenswrapper[4799]: I1129 05:00:02.938810 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406540-xd5kj" Nov 29 05:00:03 crc kubenswrapper[4799]: I1129 05:00:03.088601 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmmkt\" (UniqueName: \"kubernetes.io/projected/e3ff8087-614c-409e-8238-384a44cdc990-kube-api-access-zmmkt\") pod \"e3ff8087-614c-409e-8238-384a44cdc990\" (UID: \"e3ff8087-614c-409e-8238-384a44cdc990\") " Nov 29 05:00:03 crc kubenswrapper[4799]: I1129 05:00:03.088757 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e3ff8087-614c-409e-8238-384a44cdc990-config-volume\") pod \"e3ff8087-614c-409e-8238-384a44cdc990\" (UID: \"e3ff8087-614c-409e-8238-384a44cdc990\") " Nov 29 05:00:03 crc kubenswrapper[4799]: I1129 05:00:03.088932 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e3ff8087-614c-409e-8238-384a44cdc990-secret-volume\") pod \"e3ff8087-614c-409e-8238-384a44cdc990\" (UID: \"e3ff8087-614c-409e-8238-384a44cdc990\") " Nov 29 05:00:03 crc kubenswrapper[4799]: I1129 05:00:03.090281 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3ff8087-614c-409e-8238-384a44cdc990-config-volume" (OuterVolumeSpecName: "config-volume") pod "e3ff8087-614c-409e-8238-384a44cdc990" (UID: "e3ff8087-614c-409e-8238-384a44cdc990"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 05:00:03 crc kubenswrapper[4799]: I1129 05:00:03.095757 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3ff8087-614c-409e-8238-384a44cdc990-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e3ff8087-614c-409e-8238-384a44cdc990" (UID: "e3ff8087-614c-409e-8238-384a44cdc990"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:00:03 crc kubenswrapper[4799]: I1129 05:00:03.101949 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3ff8087-614c-409e-8238-384a44cdc990-kube-api-access-zmmkt" (OuterVolumeSpecName: "kube-api-access-zmmkt") pod "e3ff8087-614c-409e-8238-384a44cdc990" (UID: "e3ff8087-614c-409e-8238-384a44cdc990"). InnerVolumeSpecName "kube-api-access-zmmkt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:00:03 crc kubenswrapper[4799]: I1129 05:00:03.191849 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmmkt\" (UniqueName: \"kubernetes.io/projected/e3ff8087-614c-409e-8238-384a44cdc990-kube-api-access-zmmkt\") on node \"crc\" DevicePath \"\"" Nov 29 05:00:03 crc kubenswrapper[4799]: I1129 05:00:03.191921 4799 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e3ff8087-614c-409e-8238-384a44cdc990-config-volume\") on node \"crc\" DevicePath \"\"" Nov 29 05:00:03 crc kubenswrapper[4799]: I1129 05:00:03.191944 4799 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e3ff8087-614c-409e-8238-384a44cdc990-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 29 05:00:03 crc kubenswrapper[4799]: I1129 05:00:03.648245 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29406540-xd5kj" event={"ID":"e3ff8087-614c-409e-8238-384a44cdc990","Type":"ContainerDied","Data":"f3220f9d699cfc22bf09744d5e28fa265bd01aa71a93813052adba6f7160de69"} Nov 29 05:00:03 crc kubenswrapper[4799]: I1129 05:00:03.648308 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f3220f9d699cfc22bf09744d5e28fa265bd01aa71a93813052adba6f7160de69" Nov 29 05:00:03 crc kubenswrapper[4799]: I1129 05:00:03.648345 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406540-xd5kj" Nov 29 05:00:37 crc kubenswrapper[4799]: I1129 05:00:37.498013 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 05:00:37 crc kubenswrapper[4799]: I1129 05:00:37.498674 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 05:00:49 crc kubenswrapper[4799]: I1129 05:00:49.525016 4799 scope.go:117] "RemoveContainer" containerID="d31ed74a4ffe842e5ce4e12afae41c8602e5381cb90d93d341a7b3e77a97a520" Nov 29 05:01:00 crc kubenswrapper[4799]: I1129 05:01:00.173933 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29406541-dvcxf"] Nov 29 05:01:00 crc kubenswrapper[4799]: E1129 05:01:00.175697 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3ff8087-614c-409e-8238-384a44cdc990" containerName="collect-profiles" Nov 29 05:01:00 crc kubenswrapper[4799]: I1129 05:01:00.175892 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3ff8087-614c-409e-8238-384a44cdc990" containerName="collect-profiles" Nov 29 05:01:00 crc kubenswrapper[4799]: I1129 05:01:00.176238 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3ff8087-614c-409e-8238-384a44cdc990" containerName="collect-profiles" Nov 29 05:01:00 crc kubenswrapper[4799]: I1129 05:01:00.177254 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29406541-dvcxf" Nov 29 05:01:00 crc kubenswrapper[4799]: I1129 05:01:00.190246 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29406541-dvcxf"] Nov 29 05:01:00 crc kubenswrapper[4799]: I1129 05:01:00.232243 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b4eb06c-ec4a-45f3-b316-34a4410ea46b-config-data\") pod \"keystone-cron-29406541-dvcxf\" (UID: \"4b4eb06c-ec4a-45f3-b316-34a4410ea46b\") " pod="openstack/keystone-cron-29406541-dvcxf" Nov 29 05:01:00 crc kubenswrapper[4799]: I1129 05:01:00.232514 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4b4eb06c-ec4a-45f3-b316-34a4410ea46b-fernet-keys\") pod \"keystone-cron-29406541-dvcxf\" (UID: \"4b4eb06c-ec4a-45f3-b316-34a4410ea46b\") " pod="openstack/keystone-cron-29406541-dvcxf" Nov 29 05:01:00 crc kubenswrapper[4799]: I1129 05:01:00.232686 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b4eb06c-ec4a-45f3-b316-34a4410ea46b-combined-ca-bundle\") pod \"keystone-cron-29406541-dvcxf\" (UID: \"4b4eb06c-ec4a-45f3-b316-34a4410ea46b\") " pod="openstack/keystone-cron-29406541-dvcxf" Nov 29 05:01:00 crc kubenswrapper[4799]: I1129 05:01:00.232926 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2brx\" (UniqueName: \"kubernetes.io/projected/4b4eb06c-ec4a-45f3-b316-34a4410ea46b-kube-api-access-n2brx\") pod \"keystone-cron-29406541-dvcxf\" (UID: \"4b4eb06c-ec4a-45f3-b316-34a4410ea46b\") " pod="openstack/keystone-cron-29406541-dvcxf" Nov 29 05:01:00 crc kubenswrapper[4799]: I1129 05:01:00.334559 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2brx\" (UniqueName: \"kubernetes.io/projected/4b4eb06c-ec4a-45f3-b316-34a4410ea46b-kube-api-access-n2brx\") pod \"keystone-cron-29406541-dvcxf\" (UID: \"4b4eb06c-ec4a-45f3-b316-34a4410ea46b\") " pod="openstack/keystone-cron-29406541-dvcxf" Nov 29 05:01:00 crc kubenswrapper[4799]: I1129 05:01:00.334946 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b4eb06c-ec4a-45f3-b316-34a4410ea46b-config-data\") pod \"keystone-cron-29406541-dvcxf\" (UID: \"4b4eb06c-ec4a-45f3-b316-34a4410ea46b\") " pod="openstack/keystone-cron-29406541-dvcxf" Nov 29 05:01:00 crc kubenswrapper[4799]: I1129 05:01:00.335040 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4b4eb06c-ec4a-45f3-b316-34a4410ea46b-fernet-keys\") pod \"keystone-cron-29406541-dvcxf\" (UID: \"4b4eb06c-ec4a-45f3-b316-34a4410ea46b\") " pod="openstack/keystone-cron-29406541-dvcxf" Nov 29 05:01:00 crc kubenswrapper[4799]: I1129 05:01:00.335125 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b4eb06c-ec4a-45f3-b316-34a4410ea46b-combined-ca-bundle\") pod \"keystone-cron-29406541-dvcxf\" (UID: \"4b4eb06c-ec4a-45f3-b316-34a4410ea46b\") " pod="openstack/keystone-cron-29406541-dvcxf" Nov 29 05:01:00 crc kubenswrapper[4799]: I1129 05:01:00.353352 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4b4eb06c-ec4a-45f3-b316-34a4410ea46b-fernet-keys\") pod \"keystone-cron-29406541-dvcxf\" (UID: \"4b4eb06c-ec4a-45f3-b316-34a4410ea46b\") " pod="openstack/keystone-cron-29406541-dvcxf" Nov 29 05:01:00 crc kubenswrapper[4799]: I1129 05:01:00.353486 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b4eb06c-ec4a-45f3-b316-34a4410ea46b-config-data\") pod \"keystone-cron-29406541-dvcxf\" (UID: \"4b4eb06c-ec4a-45f3-b316-34a4410ea46b\") " pod="openstack/keystone-cron-29406541-dvcxf" Nov 29 05:01:00 crc kubenswrapper[4799]: I1129 05:01:00.353648 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b4eb06c-ec4a-45f3-b316-34a4410ea46b-combined-ca-bundle\") pod \"keystone-cron-29406541-dvcxf\" (UID: \"4b4eb06c-ec4a-45f3-b316-34a4410ea46b\") " pod="openstack/keystone-cron-29406541-dvcxf" Nov 29 05:01:00 crc kubenswrapper[4799]: I1129 05:01:00.362998 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2brx\" (UniqueName: \"kubernetes.io/projected/4b4eb06c-ec4a-45f3-b316-34a4410ea46b-kube-api-access-n2brx\") pod \"keystone-cron-29406541-dvcxf\" (UID: \"4b4eb06c-ec4a-45f3-b316-34a4410ea46b\") " pod="openstack/keystone-cron-29406541-dvcxf" Nov 29 05:01:00 crc kubenswrapper[4799]: I1129 05:01:00.502717 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29406541-dvcxf" Nov 29 05:01:01 crc kubenswrapper[4799]: I1129 05:01:01.014577 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29406541-dvcxf"] Nov 29 05:01:01 crc kubenswrapper[4799]: I1129 05:01:01.360780 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29406541-dvcxf" event={"ID":"4b4eb06c-ec4a-45f3-b316-34a4410ea46b","Type":"ContainerStarted","Data":"6d844d4f78f76af0d612d05f2628b48344f934d74e26ea9c185f4a8fe0b2ed9e"} Nov 29 05:01:01 crc kubenswrapper[4799]: I1129 05:01:01.361230 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29406541-dvcxf" event={"ID":"4b4eb06c-ec4a-45f3-b316-34a4410ea46b","Type":"ContainerStarted","Data":"46012c5f56aaca05e542505d0b17925ed6fb6b195160090f8d11613f00bc4bc0"} Nov 29 05:01:01 crc kubenswrapper[4799]: I1129 05:01:01.394290 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29406541-dvcxf" podStartSLOduration=1.3942650269999999 podStartE2EDuration="1.394265027s" podCreationTimestamp="2025-11-29 05:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 05:01:01.377599969 +0000 UTC m=+1337.020530369" watchObservedRunningTime="2025-11-29 05:01:01.394265027 +0000 UTC m=+1337.037195437" Nov 29 05:01:03 crc kubenswrapper[4799]: I1129 05:01:03.385050 4799 generic.go:334] "Generic (PLEG): container finished" podID="4b4eb06c-ec4a-45f3-b316-34a4410ea46b" containerID="6d844d4f78f76af0d612d05f2628b48344f934d74e26ea9c185f4a8fe0b2ed9e" exitCode=0 Nov 29 05:01:03 crc kubenswrapper[4799]: I1129 05:01:03.385122 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29406541-dvcxf" event={"ID":"4b4eb06c-ec4a-45f3-b316-34a4410ea46b","Type":"ContainerDied","Data":"6d844d4f78f76af0d612d05f2628b48344f934d74e26ea9c185f4a8fe0b2ed9e"} Nov 29 05:01:04 crc kubenswrapper[4799]: I1129 05:01:04.826003 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29406541-dvcxf" Nov 29 05:01:04 crc kubenswrapper[4799]: I1129 05:01:04.939061 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4b4eb06c-ec4a-45f3-b316-34a4410ea46b-fernet-keys\") pod \"4b4eb06c-ec4a-45f3-b316-34a4410ea46b\" (UID: \"4b4eb06c-ec4a-45f3-b316-34a4410ea46b\") " Nov 29 05:01:04 crc kubenswrapper[4799]: I1129 05:01:04.939597 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b4eb06c-ec4a-45f3-b316-34a4410ea46b-config-data\") pod \"4b4eb06c-ec4a-45f3-b316-34a4410ea46b\" (UID: \"4b4eb06c-ec4a-45f3-b316-34a4410ea46b\") " Nov 29 05:01:04 crc kubenswrapper[4799]: I1129 05:01:04.939643 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b4eb06c-ec4a-45f3-b316-34a4410ea46b-combined-ca-bundle\") pod \"4b4eb06c-ec4a-45f3-b316-34a4410ea46b\" (UID: \"4b4eb06c-ec4a-45f3-b316-34a4410ea46b\") " Nov 29 05:01:04 crc kubenswrapper[4799]: I1129 05:01:04.939682 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n2brx\" (UniqueName: \"kubernetes.io/projected/4b4eb06c-ec4a-45f3-b316-34a4410ea46b-kube-api-access-n2brx\") pod \"4b4eb06c-ec4a-45f3-b316-34a4410ea46b\" (UID: \"4b4eb06c-ec4a-45f3-b316-34a4410ea46b\") " Nov 29 05:01:04 crc kubenswrapper[4799]: I1129 05:01:04.946987 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b4eb06c-ec4a-45f3-b316-34a4410ea46b-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "4b4eb06c-ec4a-45f3-b316-34a4410ea46b" (UID: "4b4eb06c-ec4a-45f3-b316-34a4410ea46b"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:01:04 crc kubenswrapper[4799]: I1129 05:01:04.948725 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b4eb06c-ec4a-45f3-b316-34a4410ea46b-kube-api-access-n2brx" (OuterVolumeSpecName: "kube-api-access-n2brx") pod "4b4eb06c-ec4a-45f3-b316-34a4410ea46b" (UID: "4b4eb06c-ec4a-45f3-b316-34a4410ea46b"). InnerVolumeSpecName "kube-api-access-n2brx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:01:04 crc kubenswrapper[4799]: I1129 05:01:04.971153 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b4eb06c-ec4a-45f3-b316-34a4410ea46b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4b4eb06c-ec4a-45f3-b316-34a4410ea46b" (UID: "4b4eb06c-ec4a-45f3-b316-34a4410ea46b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:01:05 crc kubenswrapper[4799]: I1129 05:01:05.009831 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b4eb06c-ec4a-45f3-b316-34a4410ea46b-config-data" (OuterVolumeSpecName: "config-data") pod "4b4eb06c-ec4a-45f3-b316-34a4410ea46b" (UID: "4b4eb06c-ec4a-45f3-b316-34a4410ea46b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:01:05 crc kubenswrapper[4799]: I1129 05:01:05.041977 4799 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4b4eb06c-ec4a-45f3-b316-34a4410ea46b-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 29 05:01:05 crc kubenswrapper[4799]: I1129 05:01:05.042179 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b4eb06c-ec4a-45f3-b316-34a4410ea46b-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 05:01:05 crc kubenswrapper[4799]: I1129 05:01:05.042238 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b4eb06c-ec4a-45f3-b316-34a4410ea46b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 05:01:05 crc kubenswrapper[4799]: I1129 05:01:05.042336 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n2brx\" (UniqueName: \"kubernetes.io/projected/4b4eb06c-ec4a-45f3-b316-34a4410ea46b-kube-api-access-n2brx\") on node \"crc\" DevicePath \"\"" Nov 29 05:01:05 crc kubenswrapper[4799]: I1129 05:01:05.418825 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29406541-dvcxf" event={"ID":"4b4eb06c-ec4a-45f3-b316-34a4410ea46b","Type":"ContainerDied","Data":"46012c5f56aaca05e542505d0b17925ed6fb6b195160090f8d11613f00bc4bc0"} Nov 29 05:01:05 crc kubenswrapper[4799]: I1129 05:01:05.418890 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="46012c5f56aaca05e542505d0b17925ed6fb6b195160090f8d11613f00bc4bc0" Nov 29 05:01:05 crc kubenswrapper[4799]: I1129 05:01:05.419482 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29406541-dvcxf" Nov 29 05:01:07 crc kubenswrapper[4799]: I1129 05:01:07.497214 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 05:01:07 crc kubenswrapper[4799]: I1129 05:01:07.497615 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 05:01:37 crc kubenswrapper[4799]: I1129 05:01:37.497597 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 05:01:37 crc kubenswrapper[4799]: I1129 05:01:37.498833 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 05:01:37 crc kubenswrapper[4799]: I1129 05:01:37.499115 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-26t88" Nov 29 05:01:37 crc kubenswrapper[4799]: I1129 05:01:37.500857 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1b0c3452acd5fda0514988bec8ce2870ba0271b8c8c2d6b63a0f1b06bb0db0a1"} pod="openshift-machine-config-operator/machine-config-daemon-26t88" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 29 05:01:37 crc kubenswrapper[4799]: I1129 05:01:37.501022 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" containerID="cri-o://1b0c3452acd5fda0514988bec8ce2870ba0271b8c8c2d6b63a0f1b06bb0db0a1" gracePeriod=600 Nov 29 05:01:37 crc kubenswrapper[4799]: I1129 05:01:37.882962 4799 generic.go:334] "Generic (PLEG): container finished" podID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerID="1b0c3452acd5fda0514988bec8ce2870ba0271b8c8c2d6b63a0f1b06bb0db0a1" exitCode=0 Nov 29 05:01:37 crc kubenswrapper[4799]: I1129 05:01:37.883058 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" event={"ID":"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8","Type":"ContainerDied","Data":"1b0c3452acd5fda0514988bec8ce2870ba0271b8c8c2d6b63a0f1b06bb0db0a1"} Nov 29 05:01:37 crc kubenswrapper[4799]: I1129 05:01:37.883666 4799 scope.go:117] "RemoveContainer" containerID="0bb5e3708c6fb82d15eb8dc7170b3147f6cdc27661c141f7e521f399422452a4" Nov 29 05:01:38 crc kubenswrapper[4799]: I1129 05:01:38.899045 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" event={"ID":"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8","Type":"ContainerStarted","Data":"1724fe9a3398c865ba5cfd6fe24e206283bd68589be6152c56813bf9cd943125"} Nov 29 05:01:49 crc kubenswrapper[4799]: I1129 05:01:49.613661 4799 scope.go:117] "RemoveContainer" containerID="af15be66d22796a789def34857a97e75d27b749e255595a4c9a405ba71d3db05" Nov 29 05:01:49 crc kubenswrapper[4799]: I1129 05:01:49.653520 4799 scope.go:117] "RemoveContainer" containerID="0206e122fb25f53bced3ad79d205461cb91bdecdbdf24990b6664f0037824b4c" Nov 29 05:01:49 crc kubenswrapper[4799]: I1129 05:01:49.673655 4799 scope.go:117] "RemoveContainer" containerID="9bb14bd0f51cef201cbf42d18d47d64600f1916ddc5961f149e18335ea33cc59" Nov 29 05:02:51 crc kubenswrapper[4799]: I1129 05:02:51.767746 4799 generic.go:334] "Generic (PLEG): container finished" podID="2e328538-a1ef-49dd-9ecd-d83560c36e15" containerID="5d2ced6ff32786bd7820e5889d3ac4340084b3f818960ca1cd8507dbe9a6b4a6" exitCode=0 Nov 29 05:02:51 crc kubenswrapper[4799]: I1129 05:02:51.767870 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pvn8g" event={"ID":"2e328538-a1ef-49dd-9ecd-d83560c36e15","Type":"ContainerDied","Data":"5d2ced6ff32786bd7820e5889d3ac4340084b3f818960ca1cd8507dbe9a6b4a6"} Nov 29 05:02:53 crc kubenswrapper[4799]: I1129 05:02:53.296618 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pvn8g" Nov 29 05:02:53 crc kubenswrapper[4799]: I1129 05:02:53.404522 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e328538-a1ef-49dd-9ecd-d83560c36e15-bootstrap-combined-ca-bundle\") pod \"2e328538-a1ef-49dd-9ecd-d83560c36e15\" (UID: \"2e328538-a1ef-49dd-9ecd-d83560c36e15\") " Nov 29 05:02:53 crc kubenswrapper[4799]: I1129 05:02:53.404685 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2e328538-a1ef-49dd-9ecd-d83560c36e15-ssh-key\") pod \"2e328538-a1ef-49dd-9ecd-d83560c36e15\" (UID: \"2e328538-a1ef-49dd-9ecd-d83560c36e15\") " Nov 29 05:02:53 crc kubenswrapper[4799]: I1129 05:02:53.404867 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wr24r\" (UniqueName: \"kubernetes.io/projected/2e328538-a1ef-49dd-9ecd-d83560c36e15-kube-api-access-wr24r\") pod \"2e328538-a1ef-49dd-9ecd-d83560c36e15\" (UID: \"2e328538-a1ef-49dd-9ecd-d83560c36e15\") " Nov 29 05:02:53 crc kubenswrapper[4799]: I1129 05:02:53.404968 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2e328538-a1ef-49dd-9ecd-d83560c36e15-inventory\") pod \"2e328538-a1ef-49dd-9ecd-d83560c36e15\" (UID: \"2e328538-a1ef-49dd-9ecd-d83560c36e15\") " Nov 29 05:02:53 crc kubenswrapper[4799]: I1129 05:02:53.418204 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e328538-a1ef-49dd-9ecd-d83560c36e15-kube-api-access-wr24r" (OuterVolumeSpecName: "kube-api-access-wr24r") pod "2e328538-a1ef-49dd-9ecd-d83560c36e15" (UID: "2e328538-a1ef-49dd-9ecd-d83560c36e15"). InnerVolumeSpecName "kube-api-access-wr24r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:02:53 crc kubenswrapper[4799]: I1129 05:02:53.418308 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e328538-a1ef-49dd-9ecd-d83560c36e15-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "2e328538-a1ef-49dd-9ecd-d83560c36e15" (UID: "2e328538-a1ef-49dd-9ecd-d83560c36e15"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:02:53 crc kubenswrapper[4799]: I1129 05:02:53.441650 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e328538-a1ef-49dd-9ecd-d83560c36e15-inventory" (OuterVolumeSpecName: "inventory") pod "2e328538-a1ef-49dd-9ecd-d83560c36e15" (UID: "2e328538-a1ef-49dd-9ecd-d83560c36e15"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:02:53 crc kubenswrapper[4799]: I1129 05:02:53.448982 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e328538-a1ef-49dd-9ecd-d83560c36e15-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2e328538-a1ef-49dd-9ecd-d83560c36e15" (UID: "2e328538-a1ef-49dd-9ecd-d83560c36e15"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:02:53 crc kubenswrapper[4799]: I1129 05:02:53.507910 4799 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e328538-a1ef-49dd-9ecd-d83560c36e15-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 05:02:53 crc kubenswrapper[4799]: I1129 05:02:53.507977 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2e328538-a1ef-49dd-9ecd-d83560c36e15-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 29 05:02:53 crc kubenswrapper[4799]: I1129 05:02:53.507994 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wr24r\" (UniqueName: \"kubernetes.io/projected/2e328538-a1ef-49dd-9ecd-d83560c36e15-kube-api-access-wr24r\") on node \"crc\" DevicePath \"\"" Nov 29 05:02:53 crc kubenswrapper[4799]: I1129 05:02:53.508007 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2e328538-a1ef-49dd-9ecd-d83560c36e15-inventory\") on node \"crc\" DevicePath \"\"" Nov 29 05:02:53 crc kubenswrapper[4799]: I1129 05:02:53.797663 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pvn8g" event={"ID":"2e328538-a1ef-49dd-9ecd-d83560c36e15","Type":"ContainerDied","Data":"76b5755384e79b20960968a191b64ce8615ea12ab98c2bc8a52c014ade4a78ab"} Nov 29 05:02:53 crc kubenswrapper[4799]: I1129 05:02:53.797736 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="76b5755384e79b20960968a191b64ce8615ea12ab98c2bc8a52c014ade4a78ab" Nov 29 05:02:53 crc kubenswrapper[4799]: I1129 05:02:53.798032 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pvn8g" Nov 29 05:02:53 crc kubenswrapper[4799]: I1129 05:02:53.903959 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-whnxl"] Nov 29 05:02:53 crc kubenswrapper[4799]: E1129 05:02:53.904580 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b4eb06c-ec4a-45f3-b316-34a4410ea46b" containerName="keystone-cron" Nov 29 05:02:53 crc kubenswrapper[4799]: I1129 05:02:53.904605 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b4eb06c-ec4a-45f3-b316-34a4410ea46b" containerName="keystone-cron" Nov 29 05:02:53 crc kubenswrapper[4799]: E1129 05:02:53.904646 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e328538-a1ef-49dd-9ecd-d83560c36e15" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 29 05:02:53 crc kubenswrapper[4799]: I1129 05:02:53.904656 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e328538-a1ef-49dd-9ecd-d83560c36e15" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 29 05:02:53 crc kubenswrapper[4799]: I1129 05:02:53.904885 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b4eb06c-ec4a-45f3-b316-34a4410ea46b" containerName="keystone-cron" Nov 29 05:02:53 crc kubenswrapper[4799]: I1129 05:02:53.904912 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e328538-a1ef-49dd-9ecd-d83560c36e15" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 29 05:02:53 crc kubenswrapper[4799]: I1129 05:02:53.905785 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-whnxl" Nov 29 05:02:53 crc kubenswrapper[4799]: I1129 05:02:53.909304 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 29 05:02:53 crc kubenswrapper[4799]: I1129 05:02:53.909377 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 29 05:02:53 crc kubenswrapper[4799]: I1129 05:02:53.909394 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 29 05:02:53 crc kubenswrapper[4799]: I1129 05:02:53.910732 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ljwtk" Nov 29 05:02:53 crc kubenswrapper[4799]: I1129 05:02:53.919175 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-whnxl"] Nov 29 05:02:54 crc kubenswrapper[4799]: I1129 05:02:54.038617 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f71ef031-f30f-45aa-9d42-eb3981250587-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-whnxl\" (UID: \"f71ef031-f30f-45aa-9d42-eb3981250587\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-whnxl" Nov 29 05:02:54 crc kubenswrapper[4799]: I1129 05:02:54.038744 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f71ef031-f30f-45aa-9d42-eb3981250587-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-whnxl\" (UID: \"f71ef031-f30f-45aa-9d42-eb3981250587\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-whnxl" Nov 29 05:02:54 crc kubenswrapper[4799]: I1129 05:02:54.038857 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmxtj\" (UniqueName: \"kubernetes.io/projected/f71ef031-f30f-45aa-9d42-eb3981250587-kube-api-access-fmxtj\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-whnxl\" (UID: \"f71ef031-f30f-45aa-9d42-eb3981250587\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-whnxl" Nov 29 05:02:54 crc kubenswrapper[4799]: I1129 05:02:54.140523 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f71ef031-f30f-45aa-9d42-eb3981250587-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-whnxl\" (UID: \"f71ef031-f30f-45aa-9d42-eb3981250587\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-whnxl" Nov 29 05:02:54 crc kubenswrapper[4799]: I1129 05:02:54.141080 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f71ef031-f30f-45aa-9d42-eb3981250587-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-whnxl\" (UID: \"f71ef031-f30f-45aa-9d42-eb3981250587\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-whnxl" Nov 29 05:02:54 crc kubenswrapper[4799]: I1129 05:02:54.141401 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmxtj\" (UniqueName: \"kubernetes.io/projected/f71ef031-f30f-45aa-9d42-eb3981250587-kube-api-access-fmxtj\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-whnxl\" (UID: \"f71ef031-f30f-45aa-9d42-eb3981250587\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-whnxl" Nov 29 05:02:54 crc kubenswrapper[4799]: I1129 05:02:54.145957 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f71ef031-f30f-45aa-9d42-eb3981250587-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-whnxl\" (UID: \"f71ef031-f30f-45aa-9d42-eb3981250587\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-whnxl" Nov 29 05:02:54 crc kubenswrapper[4799]: I1129 05:02:54.148184 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f71ef031-f30f-45aa-9d42-eb3981250587-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-whnxl\" (UID: \"f71ef031-f30f-45aa-9d42-eb3981250587\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-whnxl" Nov 29 05:02:54 crc kubenswrapper[4799]: I1129 05:02:54.164024 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmxtj\" (UniqueName: \"kubernetes.io/projected/f71ef031-f30f-45aa-9d42-eb3981250587-kube-api-access-fmxtj\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-whnxl\" (UID: \"f71ef031-f30f-45aa-9d42-eb3981250587\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-whnxl" Nov 29 05:02:54 crc kubenswrapper[4799]: I1129 05:02:54.235542 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-whnxl" Nov 29 05:02:54 crc kubenswrapper[4799]: I1129 05:02:54.624814 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-whnxl"] Nov 29 05:02:54 crc kubenswrapper[4799]: I1129 05:02:54.808381 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-whnxl" event={"ID":"f71ef031-f30f-45aa-9d42-eb3981250587","Type":"ContainerStarted","Data":"895acabd163e0b12c2e3d95ee0624e2368b9a87d87c4602a7cfe7994c9f452d2"} Nov 29 05:02:55 crc kubenswrapper[4799]: I1129 05:02:55.819821 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-whnxl" event={"ID":"f71ef031-f30f-45aa-9d42-eb3981250587","Type":"ContainerStarted","Data":"f7dd788a7b11409399c220143599ae99cc4f8c536ff0e788b64946836e99500d"} Nov 29 05:02:55 crc kubenswrapper[4799]: I1129 05:02:55.847417 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-whnxl" podStartSLOduration=2.2553533359999998 podStartE2EDuration="2.847398677s" podCreationTimestamp="2025-11-29 05:02:53 +0000 UTC" firstStartedPulling="2025-11-29 05:02:54.631868264 +0000 UTC m=+1450.274798664" lastFinishedPulling="2025-11-29 05:02:55.223913615 +0000 UTC m=+1450.866844005" observedRunningTime="2025-11-29 05:02:55.839233767 +0000 UTC m=+1451.482164167" watchObservedRunningTime="2025-11-29 05:02:55.847398677 +0000 UTC m=+1451.490329077" Nov 29 05:03:37 crc kubenswrapper[4799]: I1129 05:03:37.497940 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 05:03:37 crc kubenswrapper[4799]: I1129 05:03:37.499023 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 05:03:49 crc kubenswrapper[4799]: I1129 05:03:49.792305 4799 scope.go:117] "RemoveContainer" containerID="cd2b609ab12cb328932676754756fc40228b18f8b81ffb0ef43d61eb2664f941" Nov 29 05:03:49 crc kubenswrapper[4799]: I1129 05:03:49.826326 4799 scope.go:117] "RemoveContainer" containerID="51eb9bd77778fbf6140e3e6018fb61b2ac6ff0ca188c70a433822c521d2a1908" Nov 29 05:04:00 crc kubenswrapper[4799]: I1129 05:04:00.951215 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-sw7jv"] Nov 29 05:04:00 crc kubenswrapper[4799]: I1129 05:04:00.955399 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sw7jv" Nov 29 05:04:00 crc kubenswrapper[4799]: I1129 05:04:00.964161 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sw7jv"] Nov 29 05:04:01 crc kubenswrapper[4799]: I1129 05:04:01.091962 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/524e0340-1177-4515-85cd-c1ae8ae6fd27-utilities\") pod \"community-operators-sw7jv\" (UID: \"524e0340-1177-4515-85cd-c1ae8ae6fd27\") " pod="openshift-marketplace/community-operators-sw7jv" Nov 29 05:04:01 crc kubenswrapper[4799]: I1129 05:04:01.092336 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxn2k\" (UniqueName: \"kubernetes.io/projected/524e0340-1177-4515-85cd-c1ae8ae6fd27-kube-api-access-xxn2k\") pod \"community-operators-sw7jv\" (UID: \"524e0340-1177-4515-85cd-c1ae8ae6fd27\") " pod="openshift-marketplace/community-operators-sw7jv" Nov 29 05:04:01 crc kubenswrapper[4799]: I1129 05:04:01.092431 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/524e0340-1177-4515-85cd-c1ae8ae6fd27-catalog-content\") pod \"community-operators-sw7jv\" (UID: \"524e0340-1177-4515-85cd-c1ae8ae6fd27\") " pod="openshift-marketplace/community-operators-sw7jv" Nov 29 05:04:01 crc kubenswrapper[4799]: I1129 05:04:01.195509 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxn2k\" (UniqueName: \"kubernetes.io/projected/524e0340-1177-4515-85cd-c1ae8ae6fd27-kube-api-access-xxn2k\") pod \"community-operators-sw7jv\" (UID: \"524e0340-1177-4515-85cd-c1ae8ae6fd27\") " pod="openshift-marketplace/community-operators-sw7jv" Nov 29 05:04:01 crc kubenswrapper[4799]: I1129 05:04:01.195588 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/524e0340-1177-4515-85cd-c1ae8ae6fd27-catalog-content\") pod \"community-operators-sw7jv\" (UID: \"524e0340-1177-4515-85cd-c1ae8ae6fd27\") " pod="openshift-marketplace/community-operators-sw7jv" Nov 29 05:04:01 crc kubenswrapper[4799]: I1129 05:04:01.195744 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/524e0340-1177-4515-85cd-c1ae8ae6fd27-utilities\") pod \"community-operators-sw7jv\" (UID: \"524e0340-1177-4515-85cd-c1ae8ae6fd27\") " pod="openshift-marketplace/community-operators-sw7jv" Nov 29 05:04:01 crc kubenswrapper[4799]: I1129 05:04:01.196397 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/524e0340-1177-4515-85cd-c1ae8ae6fd27-catalog-content\") pod \"community-operators-sw7jv\" (UID: \"524e0340-1177-4515-85cd-c1ae8ae6fd27\") " pod="openshift-marketplace/community-operators-sw7jv" Nov 29 05:04:01 crc kubenswrapper[4799]: I1129 05:04:01.196509 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/524e0340-1177-4515-85cd-c1ae8ae6fd27-utilities\") pod \"community-operators-sw7jv\" (UID: \"524e0340-1177-4515-85cd-c1ae8ae6fd27\") " pod="openshift-marketplace/community-operators-sw7jv" Nov 29 05:04:01 crc kubenswrapper[4799]: I1129 05:04:01.232750 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxn2k\" (UniqueName: \"kubernetes.io/projected/524e0340-1177-4515-85cd-c1ae8ae6fd27-kube-api-access-xxn2k\") pod \"community-operators-sw7jv\" (UID: \"524e0340-1177-4515-85cd-c1ae8ae6fd27\") " pod="openshift-marketplace/community-operators-sw7jv" Nov 29 05:04:01 crc kubenswrapper[4799]: I1129 05:04:01.291421 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sw7jv" Nov 29 05:04:01 crc kubenswrapper[4799]: I1129 05:04:01.908024 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sw7jv"] Nov 29 05:04:02 crc kubenswrapper[4799]: I1129 05:04:02.640472 4799 generic.go:334] "Generic (PLEG): container finished" podID="524e0340-1177-4515-85cd-c1ae8ae6fd27" containerID="efd8d7b77732c5a43f9e3b02821360bf6378ee19ace808ad8cb5f3be2371ffb2" exitCode=0 Nov 29 05:04:02 crc kubenswrapper[4799]: I1129 05:04:02.640522 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sw7jv" event={"ID":"524e0340-1177-4515-85cd-c1ae8ae6fd27","Type":"ContainerDied","Data":"efd8d7b77732c5a43f9e3b02821360bf6378ee19ace808ad8cb5f3be2371ffb2"} Nov 29 05:04:02 crc kubenswrapper[4799]: I1129 05:04:02.640560 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sw7jv" event={"ID":"524e0340-1177-4515-85cd-c1ae8ae6fd27","Type":"ContainerStarted","Data":"18126843b3027bccd8d936c396920166eaf185694a613a69f637733917a1dae9"} Nov 29 05:04:03 crc kubenswrapper[4799]: I1129 05:04:03.652036 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sw7jv" event={"ID":"524e0340-1177-4515-85cd-c1ae8ae6fd27","Type":"ContainerStarted","Data":"c7de328a66bc8850cd162ec09dcb5290ec381082780b1c6a1a106e1f96241b4d"} Nov 29 05:04:04 crc kubenswrapper[4799]: I1129 05:04:04.339472 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-n6kps"] Nov 29 05:04:04 crc kubenswrapper[4799]: I1129 05:04:04.343853 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n6kps" Nov 29 05:04:04 crc kubenswrapper[4799]: I1129 05:04:04.355392 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-n6kps"] Nov 29 05:04:04 crc kubenswrapper[4799]: I1129 05:04:04.408638 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bdkx\" (UniqueName: \"kubernetes.io/projected/3f20449d-2953-44a1-9393-a23522699ef1-kube-api-access-7bdkx\") pod \"certified-operators-n6kps\" (UID: \"3f20449d-2953-44a1-9393-a23522699ef1\") " pod="openshift-marketplace/certified-operators-n6kps" Nov 29 05:04:04 crc kubenswrapper[4799]: I1129 05:04:04.408735 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f20449d-2953-44a1-9393-a23522699ef1-utilities\") pod \"certified-operators-n6kps\" (UID: \"3f20449d-2953-44a1-9393-a23522699ef1\") " pod="openshift-marketplace/certified-operators-n6kps" Nov 29 05:04:04 crc kubenswrapper[4799]: I1129 05:04:04.408907 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f20449d-2953-44a1-9393-a23522699ef1-catalog-content\") pod \"certified-operators-n6kps\" (UID: \"3f20449d-2953-44a1-9393-a23522699ef1\") " pod="openshift-marketplace/certified-operators-n6kps" Nov 29 05:04:04 crc kubenswrapper[4799]: I1129 05:04:04.511062 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f20449d-2953-44a1-9393-a23522699ef1-catalog-content\") pod \"certified-operators-n6kps\" (UID: \"3f20449d-2953-44a1-9393-a23522699ef1\") " pod="openshift-marketplace/certified-operators-n6kps" Nov 29 05:04:04 crc kubenswrapper[4799]: I1129 05:04:04.511205 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bdkx\" (UniqueName: \"kubernetes.io/projected/3f20449d-2953-44a1-9393-a23522699ef1-kube-api-access-7bdkx\") pod \"certified-operators-n6kps\" (UID: \"3f20449d-2953-44a1-9393-a23522699ef1\") " pod="openshift-marketplace/certified-operators-n6kps" Nov 29 05:04:04 crc kubenswrapper[4799]: I1129 05:04:04.511245 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f20449d-2953-44a1-9393-a23522699ef1-utilities\") pod \"certified-operators-n6kps\" (UID: \"3f20449d-2953-44a1-9393-a23522699ef1\") " pod="openshift-marketplace/certified-operators-n6kps" Nov 29 05:04:04 crc kubenswrapper[4799]: I1129 05:04:04.511894 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f20449d-2953-44a1-9393-a23522699ef1-utilities\") pod \"certified-operators-n6kps\" (UID: \"3f20449d-2953-44a1-9393-a23522699ef1\") " pod="openshift-marketplace/certified-operators-n6kps" Nov 29 05:04:04 crc kubenswrapper[4799]: I1129 05:04:04.512217 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f20449d-2953-44a1-9393-a23522699ef1-catalog-content\") pod \"certified-operators-n6kps\" (UID: \"3f20449d-2953-44a1-9393-a23522699ef1\") " pod="openshift-marketplace/certified-operators-n6kps" Nov 29 05:04:04 crc kubenswrapper[4799]: I1129 05:04:04.534199 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bdkx\" (UniqueName: \"kubernetes.io/projected/3f20449d-2953-44a1-9393-a23522699ef1-kube-api-access-7bdkx\") pod \"certified-operators-n6kps\" (UID: \"3f20449d-2953-44a1-9393-a23522699ef1\") " pod="openshift-marketplace/certified-operators-n6kps" Nov 29 05:04:04 crc kubenswrapper[4799]: I1129 05:04:04.666049 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n6kps" Nov 29 05:04:04 crc kubenswrapper[4799]: I1129 05:04:04.674966 4799 generic.go:334] "Generic (PLEG): container finished" podID="524e0340-1177-4515-85cd-c1ae8ae6fd27" containerID="c7de328a66bc8850cd162ec09dcb5290ec381082780b1c6a1a106e1f96241b4d" exitCode=0 Nov 29 05:04:04 crc kubenswrapper[4799]: I1129 05:04:04.675320 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sw7jv" event={"ID":"524e0340-1177-4515-85cd-c1ae8ae6fd27","Type":"ContainerDied","Data":"c7de328a66bc8850cd162ec09dcb5290ec381082780b1c6a1a106e1f96241b4d"} Nov 29 05:04:05 crc kubenswrapper[4799]: I1129 05:04:05.321744 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-n6kps"] Nov 29 05:04:05 crc kubenswrapper[4799]: I1129 05:04:05.696654 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sw7jv" event={"ID":"524e0340-1177-4515-85cd-c1ae8ae6fd27","Type":"ContainerStarted","Data":"9abf6e31cf39799fb855c7833e70129913a6139fa0c94c0be2da8fc93e57048a"} Nov 29 05:04:05 crc kubenswrapper[4799]: I1129 05:04:05.699890 4799 generic.go:334] "Generic (PLEG): container finished" podID="3f20449d-2953-44a1-9393-a23522699ef1" containerID="b0b2785eb02f6e6d414c670f896b2e9f9f3112defddebde44b7c6835e9b116dd" exitCode=0 Nov 29 05:04:05 crc kubenswrapper[4799]: I1129 05:04:05.699944 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n6kps" event={"ID":"3f20449d-2953-44a1-9393-a23522699ef1","Type":"ContainerDied","Data":"b0b2785eb02f6e6d414c670f896b2e9f9f3112defddebde44b7c6835e9b116dd"} Nov 29 05:04:05 crc kubenswrapper[4799]: I1129 05:04:05.699974 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n6kps" event={"ID":"3f20449d-2953-44a1-9393-a23522699ef1","Type":"ContainerStarted","Data":"4acc4c9cc518ae99d47adf0cbbd0b09cda1fab09a9917883ee08ce7c189a55c2"} Nov 29 05:04:05 crc kubenswrapper[4799]: I1129 05:04:05.729126 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-sw7jv" podStartSLOduration=3.1471389419999998 podStartE2EDuration="5.729106212s" podCreationTimestamp="2025-11-29 05:04:00 +0000 UTC" firstStartedPulling="2025-11-29 05:04:02.643901344 +0000 UTC m=+1518.286831744" lastFinishedPulling="2025-11-29 05:04:05.225868624 +0000 UTC m=+1520.868799014" observedRunningTime="2025-11-29 05:04:05.723349311 +0000 UTC m=+1521.366279721" watchObservedRunningTime="2025-11-29 05:04:05.729106212 +0000 UTC m=+1521.372036612" Nov 29 05:04:06 crc kubenswrapper[4799]: I1129 05:04:06.713090 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n6kps" event={"ID":"3f20449d-2953-44a1-9393-a23522699ef1","Type":"ContainerStarted","Data":"bc336017156cec0de00ed8a2cc0afdf31e53bc6f16d3f5b48eb40bb60025537e"} Nov 29 05:04:07 crc kubenswrapper[4799]: I1129 05:04:07.498319 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 05:04:07 crc kubenswrapper[4799]: I1129 05:04:07.498923 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 05:04:07 crc kubenswrapper[4799]: I1129 05:04:07.727007 4799 generic.go:334] "Generic (PLEG): container finished" podID="3f20449d-2953-44a1-9393-a23522699ef1" containerID="bc336017156cec0de00ed8a2cc0afdf31e53bc6f16d3f5b48eb40bb60025537e" exitCode=0 Nov 29 05:04:07 crc kubenswrapper[4799]: I1129 05:04:07.727065 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n6kps" event={"ID":"3f20449d-2953-44a1-9393-a23522699ef1","Type":"ContainerDied","Data":"bc336017156cec0de00ed8a2cc0afdf31e53bc6f16d3f5b48eb40bb60025537e"} Nov 29 05:04:08 crc kubenswrapper[4799]: I1129 05:04:08.737284 4799 generic.go:334] "Generic (PLEG): container finished" podID="f71ef031-f30f-45aa-9d42-eb3981250587" containerID="f7dd788a7b11409399c220143599ae99cc4f8c536ff0e788b64946836e99500d" exitCode=0 Nov 29 05:04:08 crc kubenswrapper[4799]: I1129 05:04:08.737362 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-whnxl" event={"ID":"f71ef031-f30f-45aa-9d42-eb3981250587","Type":"ContainerDied","Data":"f7dd788a7b11409399c220143599ae99cc4f8c536ff0e788b64946836e99500d"} Nov 29 05:04:08 crc kubenswrapper[4799]: I1129 05:04:08.742570 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n6kps" event={"ID":"3f20449d-2953-44a1-9393-a23522699ef1","Type":"ContainerStarted","Data":"d593881174b2a593a44cd021e6582063fcd5144169c798a928ed0490da23b37c"} Nov 29 05:04:08 crc kubenswrapper[4799]: I1129 05:04:08.786340 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-n6kps" podStartSLOduration=2.326457224 podStartE2EDuration="4.786312635s" podCreationTimestamp="2025-11-29 05:04:04 +0000 UTC" firstStartedPulling="2025-11-29 05:04:05.701837345 +0000 UTC m=+1521.344767745" lastFinishedPulling="2025-11-29 05:04:08.161692756 +0000 UTC m=+1523.804623156" observedRunningTime="2025-11-29 05:04:08.776119844 +0000 UTC m=+1524.419050244" watchObservedRunningTime="2025-11-29 05:04:08.786312635 +0000 UTC m=+1524.429243045" Nov 29 05:04:10 crc kubenswrapper[4799]: I1129 05:04:10.260928 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-whnxl" Nov 29 05:04:10 crc kubenswrapper[4799]: I1129 05:04:10.356907 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fmxtj\" (UniqueName: \"kubernetes.io/projected/f71ef031-f30f-45aa-9d42-eb3981250587-kube-api-access-fmxtj\") pod \"f71ef031-f30f-45aa-9d42-eb3981250587\" (UID: \"f71ef031-f30f-45aa-9d42-eb3981250587\") " Nov 29 05:04:10 crc kubenswrapper[4799]: I1129 05:04:10.357180 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f71ef031-f30f-45aa-9d42-eb3981250587-inventory\") pod \"f71ef031-f30f-45aa-9d42-eb3981250587\" (UID: \"f71ef031-f30f-45aa-9d42-eb3981250587\") " Nov 29 05:04:10 crc kubenswrapper[4799]: I1129 05:04:10.357222 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f71ef031-f30f-45aa-9d42-eb3981250587-ssh-key\") pod \"f71ef031-f30f-45aa-9d42-eb3981250587\" (UID: \"f71ef031-f30f-45aa-9d42-eb3981250587\") " Nov 29 05:04:10 crc kubenswrapper[4799]: I1129 05:04:10.364129 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f71ef031-f30f-45aa-9d42-eb3981250587-kube-api-access-fmxtj" (OuterVolumeSpecName: "kube-api-access-fmxtj") pod "f71ef031-f30f-45aa-9d42-eb3981250587" (UID: "f71ef031-f30f-45aa-9d42-eb3981250587"). InnerVolumeSpecName "kube-api-access-fmxtj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:04:10 crc kubenswrapper[4799]: I1129 05:04:10.390255 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f71ef031-f30f-45aa-9d42-eb3981250587-inventory" (OuterVolumeSpecName: "inventory") pod "f71ef031-f30f-45aa-9d42-eb3981250587" (UID: "f71ef031-f30f-45aa-9d42-eb3981250587"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:04:10 crc kubenswrapper[4799]: I1129 05:04:10.392569 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f71ef031-f30f-45aa-9d42-eb3981250587-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f71ef031-f30f-45aa-9d42-eb3981250587" (UID: "f71ef031-f30f-45aa-9d42-eb3981250587"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:04:10 crc kubenswrapper[4799]: I1129 05:04:10.459033 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fmxtj\" (UniqueName: \"kubernetes.io/projected/f71ef031-f30f-45aa-9d42-eb3981250587-kube-api-access-fmxtj\") on node \"crc\" DevicePath \"\"" Nov 29 05:04:10 crc kubenswrapper[4799]: I1129 05:04:10.459068 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f71ef031-f30f-45aa-9d42-eb3981250587-inventory\") on node \"crc\" DevicePath \"\"" Nov 29 05:04:10 crc kubenswrapper[4799]: I1129 05:04:10.459082 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f71ef031-f30f-45aa-9d42-eb3981250587-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 29 05:04:10 crc kubenswrapper[4799]: I1129 05:04:10.767681 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-whnxl" event={"ID":"f71ef031-f30f-45aa-9d42-eb3981250587","Type":"ContainerDied","Data":"895acabd163e0b12c2e3d95ee0624e2368b9a87d87c4602a7cfe7994c9f452d2"} Nov 29 05:04:10 crc kubenswrapper[4799]: I1129 05:04:10.767720 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-whnxl" Nov 29 05:04:10 crc kubenswrapper[4799]: I1129 05:04:10.767725 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="895acabd163e0b12c2e3d95ee0624e2368b9a87d87c4602a7cfe7994c9f452d2" Nov 29 05:04:10 crc kubenswrapper[4799]: I1129 05:04:10.917114 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wbvkf"] Nov 29 05:04:10 crc kubenswrapper[4799]: E1129 05:04:10.917494 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f71ef031-f30f-45aa-9d42-eb3981250587" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 29 05:04:10 crc kubenswrapper[4799]: I1129 05:04:10.917511 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f71ef031-f30f-45aa-9d42-eb3981250587" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 29 05:04:10 crc kubenswrapper[4799]: I1129 05:04:10.917693 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="f71ef031-f30f-45aa-9d42-eb3981250587" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 29 05:04:10 crc kubenswrapper[4799]: I1129 05:04:10.918307 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wbvkf" Nov 29 05:04:10 crc kubenswrapper[4799]: I1129 05:04:10.921679 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 29 05:04:10 crc kubenswrapper[4799]: I1129 05:04:10.921872 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ljwtk" Nov 29 05:04:10 crc kubenswrapper[4799]: I1129 05:04:10.921999 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 29 05:04:10 crc kubenswrapper[4799]: I1129 05:04:10.923003 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 29 05:04:10 crc kubenswrapper[4799]: I1129 05:04:10.933600 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wbvkf"] Nov 29 05:04:10 crc kubenswrapper[4799]: I1129 05:04:10.971216 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43fb992c-1708-4809-9c2b-db92bab1a7e7-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wbvkf\" (UID: \"43fb992c-1708-4809-9c2b-db92bab1a7e7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wbvkf" Nov 29 05:04:10 crc kubenswrapper[4799]: I1129 05:04:10.971411 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnwhw\" (UniqueName: \"kubernetes.io/projected/43fb992c-1708-4809-9c2b-db92bab1a7e7-kube-api-access-vnwhw\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wbvkf\" (UID: \"43fb992c-1708-4809-9c2b-db92bab1a7e7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wbvkf" Nov 29 05:04:10 crc kubenswrapper[4799]: I1129 05:04:10.971502 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43fb992c-1708-4809-9c2b-db92bab1a7e7-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wbvkf\" (UID: \"43fb992c-1708-4809-9c2b-db92bab1a7e7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wbvkf" Nov 29 05:04:11 crc kubenswrapper[4799]: I1129 05:04:11.073076 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43fb992c-1708-4809-9c2b-db92bab1a7e7-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wbvkf\" (UID: \"43fb992c-1708-4809-9c2b-db92bab1a7e7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wbvkf" Nov 29 05:04:11 crc kubenswrapper[4799]: I1129 05:04:11.073144 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43fb992c-1708-4809-9c2b-db92bab1a7e7-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wbvkf\" (UID: \"43fb992c-1708-4809-9c2b-db92bab1a7e7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wbvkf" Nov 29 05:04:11 crc kubenswrapper[4799]: I1129 05:04:11.073265 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnwhw\" (UniqueName: \"kubernetes.io/projected/43fb992c-1708-4809-9c2b-db92bab1a7e7-kube-api-access-vnwhw\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wbvkf\" (UID: \"43fb992c-1708-4809-9c2b-db92bab1a7e7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wbvkf" Nov 29 05:04:11 crc kubenswrapper[4799]: I1129 05:04:11.077226 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43fb992c-1708-4809-9c2b-db92bab1a7e7-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wbvkf\" (UID: \"43fb992c-1708-4809-9c2b-db92bab1a7e7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wbvkf" Nov 29 05:04:11 crc kubenswrapper[4799]: I1129 05:04:11.077276 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43fb992c-1708-4809-9c2b-db92bab1a7e7-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wbvkf\" (UID: \"43fb992c-1708-4809-9c2b-db92bab1a7e7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wbvkf" Nov 29 05:04:11 crc kubenswrapper[4799]: I1129 05:04:11.089490 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnwhw\" (UniqueName: \"kubernetes.io/projected/43fb992c-1708-4809-9c2b-db92bab1a7e7-kube-api-access-vnwhw\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wbvkf\" (UID: \"43fb992c-1708-4809-9c2b-db92bab1a7e7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wbvkf" Nov 29 05:04:11 crc kubenswrapper[4799]: I1129 05:04:11.235886 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wbvkf" Nov 29 05:04:11 crc kubenswrapper[4799]: I1129 05:04:11.292857 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-sw7jv" Nov 29 05:04:11 crc kubenswrapper[4799]: I1129 05:04:11.292949 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-sw7jv" Nov 29 05:04:11 crc kubenswrapper[4799]: I1129 05:04:11.353376 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-sw7jv" Nov 29 05:04:11 crc kubenswrapper[4799]: I1129 05:04:11.758173 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wbvkf"] Nov 29 05:04:11 crc kubenswrapper[4799]: W1129 05:04:11.763042 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod43fb992c_1708_4809_9c2b_db92bab1a7e7.slice/crio-8a77366cedbbd83d48525e560938facda5a935751500d0dd7c1e0458c6543ec0 WatchSource:0}: Error finding container 8a77366cedbbd83d48525e560938facda5a935751500d0dd7c1e0458c6543ec0: Status 404 returned error can't find the container with id 8a77366cedbbd83d48525e560938facda5a935751500d0dd7c1e0458c6543ec0 Nov 29 05:04:11 crc kubenswrapper[4799]: I1129 05:04:11.782892 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wbvkf" event={"ID":"43fb992c-1708-4809-9c2b-db92bab1a7e7","Type":"ContainerStarted","Data":"8a77366cedbbd83d48525e560938facda5a935751500d0dd7c1e0458c6543ec0"} Nov 29 05:04:11 crc kubenswrapper[4799]: I1129 05:04:11.830899 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-sw7jv" Nov 29 05:04:12 crc kubenswrapper[4799]: I1129 05:04:12.551363 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sw7jv"] Nov 29 05:04:13 crc kubenswrapper[4799]: I1129 05:04:13.806221 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wbvkf" event={"ID":"43fb992c-1708-4809-9c2b-db92bab1a7e7","Type":"ContainerStarted","Data":"3032d7d0134d29195583edf79494e0e4a42f8b7784aedd8985ff8e0591f66748"} Nov 29 05:04:13 crc kubenswrapper[4799]: I1129 05:04:13.806405 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-sw7jv" podUID="524e0340-1177-4515-85cd-c1ae8ae6fd27" containerName="registry-server" containerID="cri-o://9abf6e31cf39799fb855c7833e70129913a6139fa0c94c0be2da8fc93e57048a" gracePeriod=2 Nov 29 05:04:13 crc kubenswrapper[4799]: I1129 05:04:13.841302 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wbvkf" podStartSLOduration=2.996090068 podStartE2EDuration="3.841282267s" podCreationTimestamp="2025-11-29 05:04:10 +0000 UTC" firstStartedPulling="2025-11-29 05:04:11.766718346 +0000 UTC m=+1527.409648766" lastFinishedPulling="2025-11-29 05:04:12.611910565 +0000 UTC m=+1528.254840965" observedRunningTime="2025-11-29 05:04:13.8381257 +0000 UTC m=+1529.481056100" watchObservedRunningTime="2025-11-29 05:04:13.841282267 +0000 UTC m=+1529.484212667" Nov 29 05:04:14 crc kubenswrapper[4799]: I1129 05:04:14.267209 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sw7jv" Nov 29 05:04:14 crc kubenswrapper[4799]: I1129 05:04:14.350524 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xxn2k\" (UniqueName: \"kubernetes.io/projected/524e0340-1177-4515-85cd-c1ae8ae6fd27-kube-api-access-xxn2k\") pod \"524e0340-1177-4515-85cd-c1ae8ae6fd27\" (UID: \"524e0340-1177-4515-85cd-c1ae8ae6fd27\") " Nov 29 05:04:14 crc kubenswrapper[4799]: I1129 05:04:14.350624 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/524e0340-1177-4515-85cd-c1ae8ae6fd27-catalog-content\") pod \"524e0340-1177-4515-85cd-c1ae8ae6fd27\" (UID: \"524e0340-1177-4515-85cd-c1ae8ae6fd27\") " Nov 29 05:04:14 crc kubenswrapper[4799]: I1129 05:04:14.350809 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/524e0340-1177-4515-85cd-c1ae8ae6fd27-utilities\") pod \"524e0340-1177-4515-85cd-c1ae8ae6fd27\" (UID: \"524e0340-1177-4515-85cd-c1ae8ae6fd27\") " Nov 29 05:04:14 crc kubenswrapper[4799]: I1129 05:04:14.351704 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/524e0340-1177-4515-85cd-c1ae8ae6fd27-utilities" (OuterVolumeSpecName: "utilities") pod "524e0340-1177-4515-85cd-c1ae8ae6fd27" (UID: "524e0340-1177-4515-85cd-c1ae8ae6fd27"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:04:14 crc kubenswrapper[4799]: I1129 05:04:14.357578 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/524e0340-1177-4515-85cd-c1ae8ae6fd27-kube-api-access-xxn2k" (OuterVolumeSpecName: "kube-api-access-xxn2k") pod "524e0340-1177-4515-85cd-c1ae8ae6fd27" (UID: "524e0340-1177-4515-85cd-c1ae8ae6fd27"). InnerVolumeSpecName "kube-api-access-xxn2k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:04:14 crc kubenswrapper[4799]: I1129 05:04:14.422017 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/524e0340-1177-4515-85cd-c1ae8ae6fd27-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "524e0340-1177-4515-85cd-c1ae8ae6fd27" (UID: "524e0340-1177-4515-85cd-c1ae8ae6fd27"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:04:14 crc kubenswrapper[4799]: I1129 05:04:14.453235 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xxn2k\" (UniqueName: \"kubernetes.io/projected/524e0340-1177-4515-85cd-c1ae8ae6fd27-kube-api-access-xxn2k\") on node \"crc\" DevicePath \"\"" Nov 29 05:04:14 crc kubenswrapper[4799]: I1129 05:04:14.453279 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/524e0340-1177-4515-85cd-c1ae8ae6fd27-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 05:04:14 crc kubenswrapper[4799]: I1129 05:04:14.453289 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/524e0340-1177-4515-85cd-c1ae8ae6fd27-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 05:04:14 crc kubenswrapper[4799]: I1129 05:04:14.673879 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-n6kps" Nov 29 05:04:14 crc kubenswrapper[4799]: I1129 05:04:14.673931 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-n6kps" Nov 29 05:04:14 crc kubenswrapper[4799]: I1129 05:04:14.743676 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-n6kps" Nov 29 05:04:14 crc kubenswrapper[4799]: I1129 05:04:14.819522 4799 generic.go:334] "Generic (PLEG): container finished" podID="524e0340-1177-4515-85cd-c1ae8ae6fd27" containerID="9abf6e31cf39799fb855c7833e70129913a6139fa0c94c0be2da8fc93e57048a" exitCode=0 Nov 29 05:04:14 crc kubenswrapper[4799]: I1129 05:04:14.819585 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sw7jv" Nov 29 05:04:14 crc kubenswrapper[4799]: I1129 05:04:14.819641 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sw7jv" event={"ID":"524e0340-1177-4515-85cd-c1ae8ae6fd27","Type":"ContainerDied","Data":"9abf6e31cf39799fb855c7833e70129913a6139fa0c94c0be2da8fc93e57048a"} Nov 29 05:04:14 crc kubenswrapper[4799]: I1129 05:04:14.819711 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sw7jv" event={"ID":"524e0340-1177-4515-85cd-c1ae8ae6fd27","Type":"ContainerDied","Data":"18126843b3027bccd8d936c396920166eaf185694a613a69f637733917a1dae9"} Nov 29 05:04:14 crc kubenswrapper[4799]: I1129 05:04:14.819735 4799 scope.go:117] "RemoveContainer" containerID="9abf6e31cf39799fb855c7833e70129913a6139fa0c94c0be2da8fc93e57048a" Nov 29 05:04:14 crc kubenswrapper[4799]: I1129 05:04:14.862367 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sw7jv"] Nov 29 05:04:14 crc kubenswrapper[4799]: I1129 05:04:14.873192 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-sw7jv"] Nov 29 05:04:14 crc kubenswrapper[4799]: I1129 05:04:14.873502 4799 scope.go:117] "RemoveContainer" containerID="c7de328a66bc8850cd162ec09dcb5290ec381082780b1c6a1a106e1f96241b4d" Nov 29 05:04:14 crc kubenswrapper[4799]: I1129 05:04:14.909814 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-n6kps" Nov 29 05:04:14 crc kubenswrapper[4799]: I1129 05:04:14.917866 4799 scope.go:117] "RemoveContainer" containerID="efd8d7b77732c5a43f9e3b02821360bf6378ee19ace808ad8cb5f3be2371ffb2" Nov 29 05:04:14 crc kubenswrapper[4799]: I1129 05:04:14.941420 4799 scope.go:117] "RemoveContainer" containerID="9abf6e31cf39799fb855c7833e70129913a6139fa0c94c0be2da8fc93e57048a" Nov 29 05:04:14 crc kubenswrapper[4799]: E1129 05:04:14.942012 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9abf6e31cf39799fb855c7833e70129913a6139fa0c94c0be2da8fc93e57048a\": container with ID starting with 9abf6e31cf39799fb855c7833e70129913a6139fa0c94c0be2da8fc93e57048a not found: ID does not exist" containerID="9abf6e31cf39799fb855c7833e70129913a6139fa0c94c0be2da8fc93e57048a" Nov 29 05:04:14 crc kubenswrapper[4799]: I1129 05:04:14.942067 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9abf6e31cf39799fb855c7833e70129913a6139fa0c94c0be2da8fc93e57048a"} err="failed to get container status \"9abf6e31cf39799fb855c7833e70129913a6139fa0c94c0be2da8fc93e57048a\": rpc error: code = NotFound desc = could not find container \"9abf6e31cf39799fb855c7833e70129913a6139fa0c94c0be2da8fc93e57048a\": container with ID starting with 9abf6e31cf39799fb855c7833e70129913a6139fa0c94c0be2da8fc93e57048a not found: ID does not exist" Nov 29 05:04:14 crc kubenswrapper[4799]: I1129 05:04:14.942112 4799 scope.go:117] "RemoveContainer" containerID="c7de328a66bc8850cd162ec09dcb5290ec381082780b1c6a1a106e1f96241b4d" Nov 29 05:04:14 crc kubenswrapper[4799]: E1129 05:04:14.942426 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7de328a66bc8850cd162ec09dcb5290ec381082780b1c6a1a106e1f96241b4d\": container with ID starting with c7de328a66bc8850cd162ec09dcb5290ec381082780b1c6a1a106e1f96241b4d not found: ID does not exist" containerID="c7de328a66bc8850cd162ec09dcb5290ec381082780b1c6a1a106e1f96241b4d" Nov 29 05:04:14 crc kubenswrapper[4799]: I1129 05:04:14.942457 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7de328a66bc8850cd162ec09dcb5290ec381082780b1c6a1a106e1f96241b4d"} err="failed to get container status \"c7de328a66bc8850cd162ec09dcb5290ec381082780b1c6a1a106e1f96241b4d\": rpc error: code = NotFound desc = could not find container \"c7de328a66bc8850cd162ec09dcb5290ec381082780b1c6a1a106e1f96241b4d\": container with ID starting with c7de328a66bc8850cd162ec09dcb5290ec381082780b1c6a1a106e1f96241b4d not found: ID does not exist" Nov 29 05:04:14 crc kubenswrapper[4799]: I1129 05:04:14.942473 4799 scope.go:117] "RemoveContainer" containerID="efd8d7b77732c5a43f9e3b02821360bf6378ee19ace808ad8cb5f3be2371ffb2" Nov 29 05:04:14 crc kubenswrapper[4799]: E1129 05:04:14.942830 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"efd8d7b77732c5a43f9e3b02821360bf6378ee19ace808ad8cb5f3be2371ffb2\": container with ID starting with efd8d7b77732c5a43f9e3b02821360bf6378ee19ace808ad8cb5f3be2371ffb2 not found: ID does not exist" containerID="efd8d7b77732c5a43f9e3b02821360bf6378ee19ace808ad8cb5f3be2371ffb2" Nov 29 05:04:14 crc kubenswrapper[4799]: I1129 05:04:14.942885 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"efd8d7b77732c5a43f9e3b02821360bf6378ee19ace808ad8cb5f3be2371ffb2"} err="failed to get container status \"efd8d7b77732c5a43f9e3b02821360bf6378ee19ace808ad8cb5f3be2371ffb2\": rpc error: code = NotFound desc = could not find container \"efd8d7b77732c5a43f9e3b02821360bf6378ee19ace808ad8cb5f3be2371ffb2\": container with ID starting with efd8d7b77732c5a43f9e3b02821360bf6378ee19ace808ad8cb5f3be2371ffb2 not found: ID does not exist" Nov 29 05:04:16 crc kubenswrapper[4799]: I1129 05:04:16.683542 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="524e0340-1177-4515-85cd-c1ae8ae6fd27" path="/var/lib/kubelet/pods/524e0340-1177-4515-85cd-c1ae8ae6fd27/volumes" Nov 29 05:04:16 crc kubenswrapper[4799]: I1129 05:04:16.924693 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-n6kps"] Nov 29 05:04:16 crc kubenswrapper[4799]: I1129 05:04:16.925012 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-n6kps" podUID="3f20449d-2953-44a1-9393-a23522699ef1" containerName="registry-server" containerID="cri-o://d593881174b2a593a44cd021e6582063fcd5144169c798a928ed0490da23b37c" gracePeriod=2 Nov 29 05:04:17 crc kubenswrapper[4799]: I1129 05:04:17.483703 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n6kps" Nov 29 05:04:17 crc kubenswrapper[4799]: I1129 05:04:17.530450 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f20449d-2953-44a1-9393-a23522699ef1-utilities\") pod \"3f20449d-2953-44a1-9393-a23522699ef1\" (UID: \"3f20449d-2953-44a1-9393-a23522699ef1\") " Nov 29 05:04:17 crc kubenswrapper[4799]: I1129 05:04:17.530546 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7bdkx\" (UniqueName: \"kubernetes.io/projected/3f20449d-2953-44a1-9393-a23522699ef1-kube-api-access-7bdkx\") pod \"3f20449d-2953-44a1-9393-a23522699ef1\" (UID: \"3f20449d-2953-44a1-9393-a23522699ef1\") " Nov 29 05:04:17 crc kubenswrapper[4799]: I1129 05:04:17.530604 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f20449d-2953-44a1-9393-a23522699ef1-catalog-content\") pod \"3f20449d-2953-44a1-9393-a23522699ef1\" (UID: \"3f20449d-2953-44a1-9393-a23522699ef1\") " Nov 29 05:04:17 crc kubenswrapper[4799]: I1129 05:04:17.531586 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f20449d-2953-44a1-9393-a23522699ef1-utilities" (OuterVolumeSpecName: "utilities") pod "3f20449d-2953-44a1-9393-a23522699ef1" (UID: "3f20449d-2953-44a1-9393-a23522699ef1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:04:17 crc kubenswrapper[4799]: I1129 05:04:17.544217 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f20449d-2953-44a1-9393-a23522699ef1-kube-api-access-7bdkx" (OuterVolumeSpecName: "kube-api-access-7bdkx") pod "3f20449d-2953-44a1-9393-a23522699ef1" (UID: "3f20449d-2953-44a1-9393-a23522699ef1"). InnerVolumeSpecName "kube-api-access-7bdkx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:04:17 crc kubenswrapper[4799]: I1129 05:04:17.588062 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f20449d-2953-44a1-9393-a23522699ef1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3f20449d-2953-44a1-9393-a23522699ef1" (UID: "3f20449d-2953-44a1-9393-a23522699ef1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:04:17 crc kubenswrapper[4799]: I1129 05:04:17.633983 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f20449d-2953-44a1-9393-a23522699ef1-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 05:04:17 crc kubenswrapper[4799]: I1129 05:04:17.634028 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7bdkx\" (UniqueName: \"kubernetes.io/projected/3f20449d-2953-44a1-9393-a23522699ef1-kube-api-access-7bdkx\") on node \"crc\" DevicePath \"\"" Nov 29 05:04:17 crc kubenswrapper[4799]: I1129 05:04:17.634045 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f20449d-2953-44a1-9393-a23522699ef1-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 05:04:17 crc kubenswrapper[4799]: I1129 05:04:17.866081 4799 generic.go:334] "Generic (PLEG): container finished" podID="3f20449d-2953-44a1-9393-a23522699ef1" containerID="d593881174b2a593a44cd021e6582063fcd5144169c798a928ed0490da23b37c" exitCode=0 Nov 29 05:04:17 crc kubenswrapper[4799]: I1129 05:04:17.866178 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n6kps" event={"ID":"3f20449d-2953-44a1-9393-a23522699ef1","Type":"ContainerDied","Data":"d593881174b2a593a44cd021e6582063fcd5144169c798a928ed0490da23b37c"} Nov 29 05:04:17 crc kubenswrapper[4799]: I1129 05:04:17.866680 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n6kps" event={"ID":"3f20449d-2953-44a1-9393-a23522699ef1","Type":"ContainerDied","Data":"4acc4c9cc518ae99d47adf0cbbd0b09cda1fab09a9917883ee08ce7c189a55c2"} Nov 29 05:04:17 crc kubenswrapper[4799]: I1129 05:04:17.866724 4799 scope.go:117] "RemoveContainer" containerID="d593881174b2a593a44cd021e6582063fcd5144169c798a928ed0490da23b37c" Nov 29 05:04:17 crc kubenswrapper[4799]: I1129 05:04:17.866202 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n6kps" Nov 29 05:04:17 crc kubenswrapper[4799]: I1129 05:04:17.903985 4799 scope.go:117] "RemoveContainer" containerID="bc336017156cec0de00ed8a2cc0afdf31e53bc6f16d3f5b48eb40bb60025537e" Nov 29 05:04:17 crc kubenswrapper[4799]: I1129 05:04:17.929447 4799 scope.go:117] "RemoveContainer" containerID="b0b2785eb02f6e6d414c670f896b2e9f9f3112defddebde44b7c6835e9b116dd" Nov 29 05:04:17 crc kubenswrapper[4799]: I1129 05:04:17.936549 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-n6kps"] Nov 29 05:04:17 crc kubenswrapper[4799]: I1129 05:04:17.953075 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-n6kps"] Nov 29 05:04:17 crc kubenswrapper[4799]: I1129 05:04:17.996520 4799 scope.go:117] "RemoveContainer" containerID="d593881174b2a593a44cd021e6582063fcd5144169c798a928ed0490da23b37c" Nov 29 05:04:17 crc kubenswrapper[4799]: E1129 05:04:17.997356 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d593881174b2a593a44cd021e6582063fcd5144169c798a928ed0490da23b37c\": container with ID starting with d593881174b2a593a44cd021e6582063fcd5144169c798a928ed0490da23b37c not found: ID does not exist" containerID="d593881174b2a593a44cd021e6582063fcd5144169c798a928ed0490da23b37c" Nov 29 05:04:17 crc kubenswrapper[4799]: I1129 05:04:17.997429 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d593881174b2a593a44cd021e6582063fcd5144169c798a928ed0490da23b37c"} err="failed to get container status \"d593881174b2a593a44cd021e6582063fcd5144169c798a928ed0490da23b37c\": rpc error: code = NotFound desc = could not find container \"d593881174b2a593a44cd021e6582063fcd5144169c798a928ed0490da23b37c\": container with ID starting with d593881174b2a593a44cd021e6582063fcd5144169c798a928ed0490da23b37c not found: ID does not exist" Nov 29 05:04:17 crc kubenswrapper[4799]: I1129 05:04:17.997481 4799 scope.go:117] "RemoveContainer" containerID="bc336017156cec0de00ed8a2cc0afdf31e53bc6f16d3f5b48eb40bb60025537e" Nov 29 05:04:17 crc kubenswrapper[4799]: E1129 05:04:17.998111 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc336017156cec0de00ed8a2cc0afdf31e53bc6f16d3f5b48eb40bb60025537e\": container with ID starting with bc336017156cec0de00ed8a2cc0afdf31e53bc6f16d3f5b48eb40bb60025537e not found: ID does not exist" containerID="bc336017156cec0de00ed8a2cc0afdf31e53bc6f16d3f5b48eb40bb60025537e" Nov 29 05:04:17 crc kubenswrapper[4799]: I1129 05:04:17.998167 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc336017156cec0de00ed8a2cc0afdf31e53bc6f16d3f5b48eb40bb60025537e"} err="failed to get container status \"bc336017156cec0de00ed8a2cc0afdf31e53bc6f16d3f5b48eb40bb60025537e\": rpc error: code = NotFound desc = could not find container \"bc336017156cec0de00ed8a2cc0afdf31e53bc6f16d3f5b48eb40bb60025537e\": container with ID starting with bc336017156cec0de00ed8a2cc0afdf31e53bc6f16d3f5b48eb40bb60025537e not found: ID does not exist" Nov 29 05:04:17 crc kubenswrapper[4799]: I1129 05:04:17.998200 4799 scope.go:117] "RemoveContainer" containerID="b0b2785eb02f6e6d414c670f896b2e9f9f3112defddebde44b7c6835e9b116dd" Nov 29 05:04:17 crc kubenswrapper[4799]: E1129 05:04:17.998638 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0b2785eb02f6e6d414c670f896b2e9f9f3112defddebde44b7c6835e9b116dd\": container with ID starting with b0b2785eb02f6e6d414c670f896b2e9f9f3112defddebde44b7c6835e9b116dd not found: ID does not exist" containerID="b0b2785eb02f6e6d414c670f896b2e9f9f3112defddebde44b7c6835e9b116dd" Nov 29 05:04:17 crc kubenswrapper[4799]: I1129 05:04:17.998712 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0b2785eb02f6e6d414c670f896b2e9f9f3112defddebde44b7c6835e9b116dd"} err="failed to get container status \"b0b2785eb02f6e6d414c670f896b2e9f9f3112defddebde44b7c6835e9b116dd\": rpc error: code = NotFound desc = could not find container \"b0b2785eb02f6e6d414c670f896b2e9f9f3112defddebde44b7c6835e9b116dd\": container with ID starting with b0b2785eb02f6e6d414c670f896b2e9f9f3112defddebde44b7c6835e9b116dd not found: ID does not exist" Nov 29 05:04:18 crc kubenswrapper[4799]: I1129 05:04:18.672481 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f20449d-2953-44a1-9393-a23522699ef1" path="/var/lib/kubelet/pods/3f20449d-2953-44a1-9393-a23522699ef1/volumes" Nov 29 05:04:18 crc kubenswrapper[4799]: I1129 05:04:18.884252 4799 generic.go:334] "Generic (PLEG): container finished" podID="43fb992c-1708-4809-9c2b-db92bab1a7e7" containerID="3032d7d0134d29195583edf79494e0e4a42f8b7784aedd8985ff8e0591f66748" exitCode=0 Nov 29 05:04:18 crc kubenswrapper[4799]: I1129 05:04:18.884317 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wbvkf" event={"ID":"43fb992c-1708-4809-9c2b-db92bab1a7e7","Type":"ContainerDied","Data":"3032d7d0134d29195583edf79494e0e4a42f8b7784aedd8985ff8e0591f66748"} Nov 29 05:04:20 crc kubenswrapper[4799]: I1129 05:04:20.529075 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wbvkf" Nov 29 05:04:20 crc kubenswrapper[4799]: I1129 05:04:20.606399 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43fb992c-1708-4809-9c2b-db92bab1a7e7-inventory\") pod \"43fb992c-1708-4809-9c2b-db92bab1a7e7\" (UID: \"43fb992c-1708-4809-9c2b-db92bab1a7e7\") " Nov 29 05:04:20 crc kubenswrapper[4799]: I1129 05:04:20.606541 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43fb992c-1708-4809-9c2b-db92bab1a7e7-ssh-key\") pod \"43fb992c-1708-4809-9c2b-db92bab1a7e7\" (UID: \"43fb992c-1708-4809-9c2b-db92bab1a7e7\") " Nov 29 05:04:20 crc kubenswrapper[4799]: I1129 05:04:20.606620 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vnwhw\" (UniqueName: \"kubernetes.io/projected/43fb992c-1708-4809-9c2b-db92bab1a7e7-kube-api-access-vnwhw\") pod \"43fb992c-1708-4809-9c2b-db92bab1a7e7\" (UID: \"43fb992c-1708-4809-9c2b-db92bab1a7e7\") " Nov 29 05:04:20 crc kubenswrapper[4799]: I1129 05:04:20.616696 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43fb992c-1708-4809-9c2b-db92bab1a7e7-kube-api-access-vnwhw" (OuterVolumeSpecName: "kube-api-access-vnwhw") pod "43fb992c-1708-4809-9c2b-db92bab1a7e7" (UID: "43fb992c-1708-4809-9c2b-db92bab1a7e7"). InnerVolumeSpecName "kube-api-access-vnwhw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:04:20 crc kubenswrapper[4799]: I1129 05:04:20.640501 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43fb992c-1708-4809-9c2b-db92bab1a7e7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "43fb992c-1708-4809-9c2b-db92bab1a7e7" (UID: "43fb992c-1708-4809-9c2b-db92bab1a7e7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:04:20 crc kubenswrapper[4799]: I1129 05:04:20.641532 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43fb992c-1708-4809-9c2b-db92bab1a7e7-inventory" (OuterVolumeSpecName: "inventory") pod "43fb992c-1708-4809-9c2b-db92bab1a7e7" (UID: "43fb992c-1708-4809-9c2b-db92bab1a7e7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:04:20 crc kubenswrapper[4799]: I1129 05:04:20.709159 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43fb992c-1708-4809-9c2b-db92bab1a7e7-inventory\") on node \"crc\" DevicePath \"\"" Nov 29 05:04:20 crc kubenswrapper[4799]: I1129 05:04:20.709193 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43fb992c-1708-4809-9c2b-db92bab1a7e7-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 29 05:04:20 crc kubenswrapper[4799]: I1129 05:04:20.709204 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vnwhw\" (UniqueName: \"kubernetes.io/projected/43fb992c-1708-4809-9c2b-db92bab1a7e7-kube-api-access-vnwhw\") on node \"crc\" DevicePath \"\"" Nov 29 05:04:20 crc kubenswrapper[4799]: I1129 05:04:20.906952 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wbvkf" event={"ID":"43fb992c-1708-4809-9c2b-db92bab1a7e7","Type":"ContainerDied","Data":"8a77366cedbbd83d48525e560938facda5a935751500d0dd7c1e0458c6543ec0"} Nov 29 05:04:20 crc kubenswrapper[4799]: I1129 05:04:20.907387 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8a77366cedbbd83d48525e560938facda5a935751500d0dd7c1e0458c6543ec0" Nov 29 05:04:20 crc kubenswrapper[4799]: I1129 05:04:20.907278 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wbvkf" Nov 29 05:04:21 crc kubenswrapper[4799]: I1129 05:04:21.042220 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-5p5jb"] Nov 29 05:04:21 crc kubenswrapper[4799]: E1129 05:04:21.042862 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="524e0340-1177-4515-85cd-c1ae8ae6fd27" containerName="extract-utilities" Nov 29 05:04:21 crc kubenswrapper[4799]: I1129 05:04:21.042887 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="524e0340-1177-4515-85cd-c1ae8ae6fd27" containerName="extract-utilities" Nov 29 05:04:21 crc kubenswrapper[4799]: E1129 05:04:21.042899 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f20449d-2953-44a1-9393-a23522699ef1" containerName="registry-server" Nov 29 05:04:21 crc kubenswrapper[4799]: I1129 05:04:21.042906 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f20449d-2953-44a1-9393-a23522699ef1" containerName="registry-server" Nov 29 05:04:21 crc kubenswrapper[4799]: E1129 05:04:21.042918 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="524e0340-1177-4515-85cd-c1ae8ae6fd27" containerName="registry-server" Nov 29 05:04:21 crc kubenswrapper[4799]: I1129 05:04:21.042926 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="524e0340-1177-4515-85cd-c1ae8ae6fd27" containerName="registry-server" Nov 29 05:04:21 crc kubenswrapper[4799]: E1129 05:04:21.042945 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f20449d-2953-44a1-9393-a23522699ef1" containerName="extract-content" Nov 29 05:04:21 crc kubenswrapper[4799]: I1129 05:04:21.042952 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f20449d-2953-44a1-9393-a23522699ef1" containerName="extract-content" Nov 29 05:04:21 crc kubenswrapper[4799]: E1129 05:04:21.042990 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f20449d-2953-44a1-9393-a23522699ef1" containerName="extract-utilities" Nov 29 05:04:21 crc kubenswrapper[4799]: I1129 05:04:21.043000 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f20449d-2953-44a1-9393-a23522699ef1" containerName="extract-utilities" Nov 29 05:04:21 crc kubenswrapper[4799]: E1129 05:04:21.043021 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43fb992c-1708-4809-9c2b-db92bab1a7e7" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 29 05:04:21 crc kubenswrapper[4799]: I1129 05:04:21.043033 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="43fb992c-1708-4809-9c2b-db92bab1a7e7" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 29 05:04:21 crc kubenswrapper[4799]: E1129 05:04:21.043057 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="524e0340-1177-4515-85cd-c1ae8ae6fd27" containerName="extract-content" Nov 29 05:04:21 crc kubenswrapper[4799]: I1129 05:04:21.043066 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="524e0340-1177-4515-85cd-c1ae8ae6fd27" containerName="extract-content" Nov 29 05:04:21 crc kubenswrapper[4799]: I1129 05:04:21.043327 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f20449d-2953-44a1-9393-a23522699ef1" containerName="registry-server" Nov 29 05:04:21 crc kubenswrapper[4799]: I1129 05:04:21.043343 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="524e0340-1177-4515-85cd-c1ae8ae6fd27" containerName="registry-server" Nov 29 05:04:21 crc kubenswrapper[4799]: I1129 05:04:21.043373 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="43fb992c-1708-4809-9c2b-db92bab1a7e7" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 29 05:04:21 crc kubenswrapper[4799]: I1129 05:04:21.044324 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5p5jb" Nov 29 05:04:21 crc kubenswrapper[4799]: I1129 05:04:21.047315 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 29 05:04:21 crc kubenswrapper[4799]: I1129 05:04:21.047600 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 29 05:04:21 crc kubenswrapper[4799]: I1129 05:04:21.047843 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 29 05:04:21 crc kubenswrapper[4799]: I1129 05:04:21.047886 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ljwtk" Nov 29 05:04:21 crc kubenswrapper[4799]: I1129 05:04:21.070425 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-5p5jb"] Nov 29 05:04:21 crc kubenswrapper[4799]: I1129 05:04:21.123702 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qv94x\" (UniqueName: \"kubernetes.io/projected/5eef8f30-b19c-47e4-9f26-eb2f386bc089-kube-api-access-qv94x\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-5p5jb\" (UID: \"5eef8f30-b19c-47e4-9f26-eb2f386bc089\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5p5jb" Nov 29 05:04:21 crc kubenswrapper[4799]: I1129 05:04:21.124159 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5eef8f30-b19c-47e4-9f26-eb2f386bc089-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-5p5jb\" (UID: \"5eef8f30-b19c-47e4-9f26-eb2f386bc089\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5p5jb" Nov 29 05:04:21 crc kubenswrapper[4799]: I1129 05:04:21.124279 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5eef8f30-b19c-47e4-9f26-eb2f386bc089-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-5p5jb\" (UID: \"5eef8f30-b19c-47e4-9f26-eb2f386bc089\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5p5jb" Nov 29 05:04:21 crc kubenswrapper[4799]: I1129 05:04:21.226345 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5eef8f30-b19c-47e4-9f26-eb2f386bc089-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-5p5jb\" (UID: \"5eef8f30-b19c-47e4-9f26-eb2f386bc089\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5p5jb" Nov 29 05:04:21 crc kubenswrapper[4799]: I1129 05:04:21.226443 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5eef8f30-b19c-47e4-9f26-eb2f386bc089-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-5p5jb\" (UID: \"5eef8f30-b19c-47e4-9f26-eb2f386bc089\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5p5jb" Nov 29 05:04:21 crc kubenswrapper[4799]: I1129 05:04:21.227287 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qv94x\" (UniqueName: \"kubernetes.io/projected/5eef8f30-b19c-47e4-9f26-eb2f386bc089-kube-api-access-qv94x\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-5p5jb\" (UID: \"5eef8f30-b19c-47e4-9f26-eb2f386bc089\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5p5jb" Nov 29 05:04:21 crc kubenswrapper[4799]: I1129 05:04:21.243984 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5eef8f30-b19c-47e4-9f26-eb2f386bc089-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-5p5jb\" (UID: \"5eef8f30-b19c-47e4-9f26-eb2f386bc089\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5p5jb" Nov 29 05:04:21 crc kubenswrapper[4799]: I1129 05:04:21.243983 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5eef8f30-b19c-47e4-9f26-eb2f386bc089-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-5p5jb\" (UID: \"5eef8f30-b19c-47e4-9f26-eb2f386bc089\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5p5jb" Nov 29 05:04:21 crc kubenswrapper[4799]: I1129 05:04:21.256089 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qv94x\" (UniqueName: \"kubernetes.io/projected/5eef8f30-b19c-47e4-9f26-eb2f386bc089-kube-api-access-qv94x\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-5p5jb\" (UID: \"5eef8f30-b19c-47e4-9f26-eb2f386bc089\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5p5jb" Nov 29 05:04:21 crc kubenswrapper[4799]: I1129 05:04:21.370904 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5p5jb" Nov 29 05:04:21 crc kubenswrapper[4799]: I1129 05:04:21.940932 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-5p5jb"] Nov 29 05:04:21 crc kubenswrapper[4799]: I1129 05:04:21.953601 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 29 05:04:22 crc kubenswrapper[4799]: I1129 05:04:22.935420 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5p5jb" event={"ID":"5eef8f30-b19c-47e4-9f26-eb2f386bc089","Type":"ContainerStarted","Data":"2d082b6293a5c8c05c0f55aec36c2335bfad32aae314d76b625580bff9247b53"} Nov 29 05:04:22 crc kubenswrapper[4799]: I1129 05:04:22.936098 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5p5jb" event={"ID":"5eef8f30-b19c-47e4-9f26-eb2f386bc089","Type":"ContainerStarted","Data":"cce35b3e652dac0da291f42bdae9973199945a8e63ce61580bf6df5e2ada3ff8"} Nov 29 05:04:22 crc kubenswrapper[4799]: I1129 05:04:22.960559 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5p5jb" podStartSLOduration=1.463319232 podStartE2EDuration="1.960533942s" podCreationTimestamp="2025-11-29 05:04:21 +0000 UTC" firstStartedPulling="2025-11-29 05:04:21.95340276 +0000 UTC m=+1537.596333160" lastFinishedPulling="2025-11-29 05:04:22.45061746 +0000 UTC m=+1538.093547870" observedRunningTime="2025-11-29 05:04:22.955907689 +0000 UTC m=+1538.598838099" watchObservedRunningTime="2025-11-29 05:04:22.960533942 +0000 UTC m=+1538.603464352" Nov 29 05:04:25 crc kubenswrapper[4799]: I1129 05:04:25.063664 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-s2hlx"] Nov 29 05:04:25 crc kubenswrapper[4799]: I1129 05:04:25.081660 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-5610-account-create-update-gwfrn"] Nov 29 05:04:25 crc kubenswrapper[4799]: I1129 05:04:25.090907 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-s2hlx"] Nov 29 05:04:25 crc kubenswrapper[4799]: I1129 05:04:25.100344 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-5610-account-create-update-gwfrn"] Nov 29 05:04:26 crc kubenswrapper[4799]: I1129 05:04:26.672644 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d6648ca-f3df-4345-a2ac-d147e7ba85e1" path="/var/lib/kubelet/pods/2d6648ca-f3df-4345-a2ac-d147e7ba85e1/volumes" Nov 29 05:04:26 crc kubenswrapper[4799]: I1129 05:04:26.674501 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f356f37d-81e1-4554-abfe-8d13412ae662" path="/var/lib/kubelet/pods/f356f37d-81e1-4554-abfe-8d13412ae662/volumes" Nov 29 05:04:29 crc kubenswrapper[4799]: I1129 05:04:29.062666 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-rsmmk"] Nov 29 05:04:29 crc kubenswrapper[4799]: I1129 05:04:29.082723 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-m5vgb"] Nov 29 05:04:29 crc kubenswrapper[4799]: I1129 05:04:29.103730 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-53da-account-create-update-xrx67"] Nov 29 05:04:29 crc kubenswrapper[4799]: I1129 05:04:29.123477 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-rsmmk"] Nov 29 05:04:29 crc kubenswrapper[4799]: I1129 05:04:29.150637 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-53da-account-create-update-xrx67"] Nov 29 05:04:29 crc kubenswrapper[4799]: I1129 05:04:29.164301 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-m5vgb"] Nov 29 05:04:30 crc kubenswrapper[4799]: I1129 05:04:30.046839 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-66ef-account-create-update-mcd96"] Nov 29 05:04:30 crc kubenswrapper[4799]: I1129 05:04:30.063165 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-66ef-account-create-update-mcd96"] Nov 29 05:04:30 crc kubenswrapper[4799]: I1129 05:04:30.694921 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58450435-f88c-4ad2-abd9-ff0c1c183527" path="/var/lib/kubelet/pods/58450435-f88c-4ad2-abd9-ff0c1c183527/volumes" Nov 29 05:04:30 crc kubenswrapper[4799]: I1129 05:04:30.695720 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74500694-328d-483c-8549-2bc90c72c811" path="/var/lib/kubelet/pods/74500694-328d-483c-8549-2bc90c72c811/volumes" Nov 29 05:04:30 crc kubenswrapper[4799]: I1129 05:04:30.696646 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e8bbbf6-43e4-4cd2-a3cd-326d3d11009f" path="/var/lib/kubelet/pods/9e8bbbf6-43e4-4cd2-a3cd-326d3d11009f/volumes" Nov 29 05:04:30 crc kubenswrapper[4799]: I1129 05:04:30.703436 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f67bb6d3-7729-467b-9bab-c3aba0a555ff" path="/var/lib/kubelet/pods/f67bb6d3-7729-467b-9bab-c3aba0a555ff/volumes" Nov 29 05:04:37 crc kubenswrapper[4799]: I1129 05:04:37.498039 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 05:04:37 crc kubenswrapper[4799]: I1129 05:04:37.498891 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 05:04:37 crc kubenswrapper[4799]: I1129 05:04:37.498946 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-26t88" Nov 29 05:04:37 crc kubenswrapper[4799]: I1129 05:04:37.499828 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1724fe9a3398c865ba5cfd6fe24e206283bd68589be6152c56813bf9cd943125"} pod="openshift-machine-config-operator/machine-config-daemon-26t88" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 29 05:04:37 crc kubenswrapper[4799]: I1129 05:04:37.499903 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" containerID="cri-o://1724fe9a3398c865ba5cfd6fe24e206283bd68589be6152c56813bf9cd943125" gracePeriod=600 Nov 29 05:04:37 crc kubenswrapper[4799]: E1129 05:04:37.629520 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:04:38 crc kubenswrapper[4799]: I1129 05:04:38.101943 4799 generic.go:334] "Generic (PLEG): container finished" podID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerID="1724fe9a3398c865ba5cfd6fe24e206283bd68589be6152c56813bf9cd943125" exitCode=0 Nov 29 05:04:38 crc kubenswrapper[4799]: I1129 05:04:38.102052 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" event={"ID":"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8","Type":"ContainerDied","Data":"1724fe9a3398c865ba5cfd6fe24e206283bd68589be6152c56813bf9cd943125"} Nov 29 05:04:38 crc kubenswrapper[4799]: I1129 05:04:38.102572 4799 scope.go:117] "RemoveContainer" containerID="1b0c3452acd5fda0514988bec8ce2870ba0271b8c8c2d6b63a0f1b06bb0db0a1" Nov 29 05:04:38 crc kubenswrapper[4799]: I1129 05:04:38.104011 4799 scope.go:117] "RemoveContainer" containerID="1724fe9a3398c865ba5cfd6fe24e206283bd68589be6152c56813bf9cd943125" Nov 29 05:04:38 crc kubenswrapper[4799]: E1129 05:04:38.104689 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:04:46 crc kubenswrapper[4799]: I1129 05:04:46.048073 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-9vlv8"] Nov 29 05:04:46 crc kubenswrapper[4799]: I1129 05:04:46.061531 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-tv4bk"] Nov 29 05:04:46 crc kubenswrapper[4799]: I1129 05:04:46.074315 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-tv4bk"] Nov 29 05:04:46 crc kubenswrapper[4799]: I1129 05:04:46.083655 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-9vlv8"] Nov 29 05:04:46 crc kubenswrapper[4799]: I1129 05:04:46.671035 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a762825-986d-4691-88bb-da6580cf83b1" path="/var/lib/kubelet/pods/1a762825-986d-4691-88bb-da6580cf83b1/volumes" Nov 29 05:04:46 crc kubenswrapper[4799]: I1129 05:04:46.671814 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc7d764a-f94f-4f3a-bf98-643a323f7cc4" path="/var/lib/kubelet/pods/fc7d764a-f94f-4f3a-bf98-643a323f7cc4/volumes" Nov 29 05:04:47 crc kubenswrapper[4799]: I1129 05:04:47.035574 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-7da7-account-create-update-wr6xq"] Nov 29 05:04:47 crc kubenswrapper[4799]: I1129 05:04:47.054274 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7451-account-create-update-r6m7w"] Nov 29 05:04:47 crc kubenswrapper[4799]: I1129 05:04:47.072968 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-2f96-account-create-update-4gdd6"] Nov 29 05:04:47 crc kubenswrapper[4799]: I1129 05:04:47.084293 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-7da7-account-create-update-wr6xq"] Nov 29 05:04:47 crc kubenswrapper[4799]: I1129 05:04:47.093428 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-7451-account-create-update-r6m7w"] Nov 29 05:04:47 crc kubenswrapper[4799]: I1129 05:04:47.105301 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-2f96-account-create-update-4gdd6"] Nov 29 05:04:48 crc kubenswrapper[4799]: I1129 05:04:48.047424 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-665zk"] Nov 29 05:04:48 crc kubenswrapper[4799]: I1129 05:04:48.065394 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-665zk"] Nov 29 05:04:48 crc kubenswrapper[4799]: I1129 05:04:48.671433 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29be7ae1-328d-4e71-9796-f054b8f4686f" path="/var/lib/kubelet/pods/29be7ae1-328d-4e71-9796-f054b8f4686f/volumes" Nov 29 05:04:48 crc kubenswrapper[4799]: I1129 05:04:48.672164 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac99f10b-e4cb-4847-a60b-69306ff5e8ca" path="/var/lib/kubelet/pods/ac99f10b-e4cb-4847-a60b-69306ff5e8ca/volumes" Nov 29 05:04:48 crc kubenswrapper[4799]: I1129 05:04:48.672729 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5536f91-0fec-4581-8a6d-18b365e7978b" path="/var/lib/kubelet/pods/b5536f91-0fec-4581-8a6d-18b365e7978b/volumes" Nov 29 05:04:48 crc kubenswrapper[4799]: I1129 05:04:48.673298 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee16aa26-5b84-447d-8a02-7a4199451cc3" path="/var/lib/kubelet/pods/ee16aa26-5b84-447d-8a02-7a4199451cc3/volumes" Nov 29 05:04:49 crc kubenswrapper[4799]: I1129 05:04:49.913015 4799 scope.go:117] "RemoveContainer" containerID="c54c73645b3442b36048196436e145256b136c8e6ff2a4ee944d7ff41c3bdcb6" Nov 29 05:04:49 crc kubenswrapper[4799]: I1129 05:04:49.944638 4799 scope.go:117] "RemoveContainer" containerID="1c498b3d5cde10e3f6ad0de28e3513ceb35283bd6a48265e31da17bb6ab8d6de" Nov 29 05:04:50 crc kubenswrapper[4799]: I1129 05:04:50.006030 4799 scope.go:117] "RemoveContainer" containerID="5c4d07e7972fb0dfc57af15e6f3f04e1dd0960514c48510b17ec18a5adb1a08c" Nov 29 05:04:50 crc kubenswrapper[4799]: I1129 05:04:50.061920 4799 scope.go:117] "RemoveContainer" containerID="15012f9018ab27ea86dbb3801f96976cbb660267030da824f8195264a2330db5" Nov 29 05:04:50 crc kubenswrapper[4799]: I1129 05:04:50.103308 4799 scope.go:117] "RemoveContainer" containerID="4fdff7a088d727ee5ebb687fd9227d1917f4ab6b2bc9a6c9827f0958c39caad0" Nov 29 05:04:50 crc kubenswrapper[4799]: I1129 05:04:50.145562 4799 scope.go:117] "RemoveContainer" containerID="12b12b9deb3bd8c8432d4bd911a69d63586a1503a1208c5f786aceee458ac305" Nov 29 05:04:50 crc kubenswrapper[4799]: I1129 05:04:50.193700 4799 scope.go:117] "RemoveContainer" containerID="34ce7b64a0457b12f9c8d2672d6716de8f796b0b907a0d2e1345c17138988137" Nov 29 05:04:50 crc kubenswrapper[4799]: I1129 05:04:50.224838 4799 scope.go:117] "RemoveContainer" containerID="8cfbc8ad7e190a8189c171f5196c34588d6cf9c2acc922e7de3910d7bdcbfb1d" Nov 29 05:04:50 crc kubenswrapper[4799]: I1129 05:04:50.267782 4799 scope.go:117] "RemoveContainer" containerID="ce9bfcee7e64c14e78367370775cfb0d18567f3e71859d42cd40a5d8bb798705" Nov 29 05:04:50 crc kubenswrapper[4799]: I1129 05:04:50.307764 4799 scope.go:117] "RemoveContainer" containerID="95a57181803d7c07e2e0c77db52741f06823e6d8f0fb7edf27b6731b7aefdbcf" Nov 29 05:04:50 crc kubenswrapper[4799]: I1129 05:04:50.348085 4799 scope.go:117] "RemoveContainer" containerID="14a27c445ae2fa52c9d8c0fbfaee89d0939bf5cbaebbd9c13eb82c428f831151" Nov 29 05:04:50 crc kubenswrapper[4799]: I1129 05:04:50.379000 4799 scope.go:117] "RemoveContainer" containerID="cd6bea142029b1fe43179aa046f1515ea6fafd02ee0f666c3c9647121a90d3ee" Nov 29 05:04:51 crc kubenswrapper[4799]: I1129 05:04:51.659663 4799 scope.go:117] "RemoveContainer" containerID="1724fe9a3398c865ba5cfd6fe24e206283bd68589be6152c56813bf9cd943125" Nov 29 05:04:51 crc kubenswrapper[4799]: E1129 05:04:51.660303 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:04:55 crc kubenswrapper[4799]: I1129 05:04:55.033899 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-88jbn"] Nov 29 05:04:55 crc kubenswrapper[4799]: I1129 05:04:55.043227 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-88jbn"] Nov 29 05:04:55 crc kubenswrapper[4799]: I1129 05:04:55.780675 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zhgdz"] Nov 29 05:04:55 crc kubenswrapper[4799]: I1129 05:04:55.783325 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zhgdz" Nov 29 05:04:55 crc kubenswrapper[4799]: I1129 05:04:55.792135 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zhgdz"] Nov 29 05:04:55 crc kubenswrapper[4799]: I1129 05:04:55.886400 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b97j2\" (UniqueName: \"kubernetes.io/projected/933e9784-668a-4ffd-bf47-2206eef87d98-kube-api-access-b97j2\") pod \"redhat-marketplace-zhgdz\" (UID: \"933e9784-668a-4ffd-bf47-2206eef87d98\") " pod="openshift-marketplace/redhat-marketplace-zhgdz" Nov 29 05:04:55 crc kubenswrapper[4799]: I1129 05:04:55.886891 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/933e9784-668a-4ffd-bf47-2206eef87d98-catalog-content\") pod \"redhat-marketplace-zhgdz\" (UID: \"933e9784-668a-4ffd-bf47-2206eef87d98\") " pod="openshift-marketplace/redhat-marketplace-zhgdz" Nov 29 05:04:55 crc kubenswrapper[4799]: I1129 05:04:55.886994 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/933e9784-668a-4ffd-bf47-2206eef87d98-utilities\") pod \"redhat-marketplace-zhgdz\" (UID: \"933e9784-668a-4ffd-bf47-2206eef87d98\") " pod="openshift-marketplace/redhat-marketplace-zhgdz" Nov 29 05:04:55 crc kubenswrapper[4799]: I1129 05:04:55.989374 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/933e9784-668a-4ffd-bf47-2206eef87d98-utilities\") pod \"redhat-marketplace-zhgdz\" (UID: \"933e9784-668a-4ffd-bf47-2206eef87d98\") " pod="openshift-marketplace/redhat-marketplace-zhgdz" Nov 29 05:04:55 crc kubenswrapper[4799]: I1129 05:04:55.989605 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b97j2\" (UniqueName: \"kubernetes.io/projected/933e9784-668a-4ffd-bf47-2206eef87d98-kube-api-access-b97j2\") pod \"redhat-marketplace-zhgdz\" (UID: \"933e9784-668a-4ffd-bf47-2206eef87d98\") " pod="openshift-marketplace/redhat-marketplace-zhgdz" Nov 29 05:04:55 crc kubenswrapper[4799]: I1129 05:04:55.989734 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/933e9784-668a-4ffd-bf47-2206eef87d98-catalog-content\") pod \"redhat-marketplace-zhgdz\" (UID: \"933e9784-668a-4ffd-bf47-2206eef87d98\") " pod="openshift-marketplace/redhat-marketplace-zhgdz" Nov 29 05:04:55 crc kubenswrapper[4799]: I1129 05:04:55.991047 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/933e9784-668a-4ffd-bf47-2206eef87d98-catalog-content\") pod \"redhat-marketplace-zhgdz\" (UID: \"933e9784-668a-4ffd-bf47-2206eef87d98\") " pod="openshift-marketplace/redhat-marketplace-zhgdz" Nov 29 05:04:55 crc kubenswrapper[4799]: I1129 05:04:55.991645 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/933e9784-668a-4ffd-bf47-2206eef87d98-utilities\") pod \"redhat-marketplace-zhgdz\" (UID: \"933e9784-668a-4ffd-bf47-2206eef87d98\") " pod="openshift-marketplace/redhat-marketplace-zhgdz" Nov 29 05:04:56 crc kubenswrapper[4799]: I1129 05:04:56.043051 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-gjbqp"] Nov 29 05:04:56 crc kubenswrapper[4799]: I1129 05:04:56.052144 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-gjbqp"] Nov 29 05:04:56 crc kubenswrapper[4799]: I1129 05:04:56.058021 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b97j2\" (UniqueName: \"kubernetes.io/projected/933e9784-668a-4ffd-bf47-2206eef87d98-kube-api-access-b97j2\") pod \"redhat-marketplace-zhgdz\" (UID: \"933e9784-668a-4ffd-bf47-2206eef87d98\") " pod="openshift-marketplace/redhat-marketplace-zhgdz" Nov 29 05:04:56 crc kubenswrapper[4799]: I1129 05:04:56.107710 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zhgdz" Nov 29 05:04:56 crc kubenswrapper[4799]: I1129 05:04:56.596662 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zhgdz"] Nov 29 05:04:56 crc kubenswrapper[4799]: I1129 05:04:56.672268 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5832499-91f5-48ca-9502-0931ae1ece9e" path="/var/lib/kubelet/pods/b5832499-91f5-48ca-9502-0931ae1ece9e/volumes" Nov 29 05:04:56 crc kubenswrapper[4799]: I1129 05:04:56.673167 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c77dc962-f9a7-40dc-aeb6-d07e9c1be588" path="/var/lib/kubelet/pods/c77dc962-f9a7-40dc-aeb6-d07e9c1be588/volumes" Nov 29 05:04:57 crc kubenswrapper[4799]: I1129 05:04:57.360730 4799 generic.go:334] "Generic (PLEG): container finished" podID="933e9784-668a-4ffd-bf47-2206eef87d98" containerID="45cd14d83de548e95d17e827d295454a494bf81f0c2c93cd3fc15be15803a574" exitCode=0 Nov 29 05:04:57 crc kubenswrapper[4799]: I1129 05:04:57.360825 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zhgdz" event={"ID":"933e9784-668a-4ffd-bf47-2206eef87d98","Type":"ContainerDied","Data":"45cd14d83de548e95d17e827d295454a494bf81f0c2c93cd3fc15be15803a574"} Nov 29 05:04:57 crc kubenswrapper[4799]: I1129 05:04:57.361226 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zhgdz" event={"ID":"933e9784-668a-4ffd-bf47-2206eef87d98","Type":"ContainerStarted","Data":"2bb6a7e5a3b6cb058fb74468c7150eede1b28bdd2e0d73d229b03d8800fd586b"} Nov 29 05:04:58 crc kubenswrapper[4799]: I1129 05:04:58.383942 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zhgdz" event={"ID":"933e9784-668a-4ffd-bf47-2206eef87d98","Type":"ContainerStarted","Data":"b251c02e36322b243d69c2407b6cb40b8d7d1c61d65ff9eb365b6aba84f5f1f4"} Nov 29 05:04:59 crc kubenswrapper[4799]: I1129 05:04:59.399927 4799 generic.go:334] "Generic (PLEG): container finished" podID="933e9784-668a-4ffd-bf47-2206eef87d98" containerID="b251c02e36322b243d69c2407b6cb40b8d7d1c61d65ff9eb365b6aba84f5f1f4" exitCode=0 Nov 29 05:04:59 crc kubenswrapper[4799]: I1129 05:04:59.399999 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zhgdz" event={"ID":"933e9784-668a-4ffd-bf47-2206eef87d98","Type":"ContainerDied","Data":"b251c02e36322b243d69c2407b6cb40b8d7d1c61d65ff9eb365b6aba84f5f1f4"} Nov 29 05:05:00 crc kubenswrapper[4799]: I1129 05:05:00.413271 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zhgdz" event={"ID":"933e9784-668a-4ffd-bf47-2206eef87d98","Type":"ContainerStarted","Data":"0b2075a4e9ea9576d7d07484cd53b3b187804ed0a1398eaedfe3c1195df09139"} Nov 29 05:05:00 crc kubenswrapper[4799]: I1129 05:05:00.443671 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zhgdz" podStartSLOduration=2.762978065 podStartE2EDuration="5.44364535s" podCreationTimestamp="2025-11-29 05:04:55 +0000 UTC" firstStartedPulling="2025-11-29 05:04:57.363109587 +0000 UTC m=+1573.006039977" lastFinishedPulling="2025-11-29 05:05:00.043776852 +0000 UTC m=+1575.686707262" observedRunningTime="2025-11-29 05:05:00.436085905 +0000 UTC m=+1576.079016315" watchObservedRunningTime="2025-11-29 05:05:00.44364535 +0000 UTC m=+1576.086575750" Nov 29 05:05:04 crc kubenswrapper[4799]: I1129 05:05:04.465587 4799 generic.go:334] "Generic (PLEG): container finished" podID="5eef8f30-b19c-47e4-9f26-eb2f386bc089" containerID="2d082b6293a5c8c05c0f55aec36c2335bfad32aae314d76b625580bff9247b53" exitCode=0 Nov 29 05:05:04 crc kubenswrapper[4799]: I1129 05:05:04.465861 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5p5jb" event={"ID":"5eef8f30-b19c-47e4-9f26-eb2f386bc089","Type":"ContainerDied","Data":"2d082b6293a5c8c05c0f55aec36c2335bfad32aae314d76b625580bff9247b53"} Nov 29 05:05:05 crc kubenswrapper[4799]: I1129 05:05:05.933770 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5p5jb" Nov 29 05:05:06 crc kubenswrapper[4799]: I1129 05:05:06.048132 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5eef8f30-b19c-47e4-9f26-eb2f386bc089-inventory\") pod \"5eef8f30-b19c-47e4-9f26-eb2f386bc089\" (UID: \"5eef8f30-b19c-47e4-9f26-eb2f386bc089\") " Nov 29 05:05:06 crc kubenswrapper[4799]: I1129 05:05:06.048675 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5eef8f30-b19c-47e4-9f26-eb2f386bc089-ssh-key\") pod \"5eef8f30-b19c-47e4-9f26-eb2f386bc089\" (UID: \"5eef8f30-b19c-47e4-9f26-eb2f386bc089\") " Nov 29 05:05:06 crc kubenswrapper[4799]: I1129 05:05:06.048713 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qv94x\" (UniqueName: \"kubernetes.io/projected/5eef8f30-b19c-47e4-9f26-eb2f386bc089-kube-api-access-qv94x\") pod \"5eef8f30-b19c-47e4-9f26-eb2f386bc089\" (UID: \"5eef8f30-b19c-47e4-9f26-eb2f386bc089\") " Nov 29 05:05:06 crc kubenswrapper[4799]: I1129 05:05:06.058517 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5eef8f30-b19c-47e4-9f26-eb2f386bc089-kube-api-access-qv94x" (OuterVolumeSpecName: "kube-api-access-qv94x") pod "5eef8f30-b19c-47e4-9f26-eb2f386bc089" (UID: "5eef8f30-b19c-47e4-9f26-eb2f386bc089"). InnerVolumeSpecName "kube-api-access-qv94x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:05:06 crc kubenswrapper[4799]: I1129 05:05:06.084853 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5eef8f30-b19c-47e4-9f26-eb2f386bc089-inventory" (OuterVolumeSpecName: "inventory") pod "5eef8f30-b19c-47e4-9f26-eb2f386bc089" (UID: "5eef8f30-b19c-47e4-9f26-eb2f386bc089"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:05:06 crc kubenswrapper[4799]: I1129 05:05:06.092528 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5eef8f30-b19c-47e4-9f26-eb2f386bc089-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5eef8f30-b19c-47e4-9f26-eb2f386bc089" (UID: "5eef8f30-b19c-47e4-9f26-eb2f386bc089"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:05:06 crc kubenswrapper[4799]: I1129 05:05:06.108359 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zhgdz" Nov 29 05:05:06 crc kubenswrapper[4799]: I1129 05:05:06.108433 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zhgdz" Nov 29 05:05:06 crc kubenswrapper[4799]: I1129 05:05:06.151469 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5eef8f30-b19c-47e4-9f26-eb2f386bc089-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 29 05:05:06 crc kubenswrapper[4799]: I1129 05:05:06.151512 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qv94x\" (UniqueName: \"kubernetes.io/projected/5eef8f30-b19c-47e4-9f26-eb2f386bc089-kube-api-access-qv94x\") on node \"crc\" DevicePath \"\"" Nov 29 05:05:06 crc kubenswrapper[4799]: I1129 05:05:06.151555 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5eef8f30-b19c-47e4-9f26-eb2f386bc089-inventory\") on node \"crc\" DevicePath \"\"" Nov 29 05:05:06 crc kubenswrapper[4799]: I1129 05:05:06.175145 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zhgdz" Nov 29 05:05:06 crc kubenswrapper[4799]: I1129 05:05:06.497540 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5p5jb" event={"ID":"5eef8f30-b19c-47e4-9f26-eb2f386bc089","Type":"ContainerDied","Data":"cce35b3e652dac0da291f42bdae9973199945a8e63ce61580bf6df5e2ada3ff8"} Nov 29 05:05:06 crc kubenswrapper[4799]: I1129 05:05:06.497610 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cce35b3e652dac0da291f42bdae9973199945a8e63ce61580bf6df5e2ada3ff8" Nov 29 05:05:06 crc kubenswrapper[4799]: I1129 05:05:06.497559 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5p5jb" Nov 29 05:05:06 crc kubenswrapper[4799]: I1129 05:05:06.590645 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d52b4"] Nov 29 05:05:06 crc kubenswrapper[4799]: E1129 05:05:06.591561 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5eef8f30-b19c-47e4-9f26-eb2f386bc089" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 29 05:05:06 crc kubenswrapper[4799]: I1129 05:05:06.591592 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5eef8f30-b19c-47e4-9f26-eb2f386bc089" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 29 05:05:06 crc kubenswrapper[4799]: I1129 05:05:06.591971 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="5eef8f30-b19c-47e4-9f26-eb2f386bc089" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 29 05:05:06 crc kubenswrapper[4799]: I1129 05:05:06.593247 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d52b4" Nov 29 05:05:06 crc kubenswrapper[4799]: I1129 05:05:06.599147 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ljwtk" Nov 29 05:05:06 crc kubenswrapper[4799]: I1129 05:05:06.599738 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 29 05:05:06 crc kubenswrapper[4799]: I1129 05:05:06.599939 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 29 05:05:06 crc kubenswrapper[4799]: I1129 05:05:06.599947 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 29 05:05:06 crc kubenswrapper[4799]: I1129 05:05:06.606617 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d52b4"] Nov 29 05:05:06 crc kubenswrapper[4799]: I1129 05:05:06.628180 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zhgdz" Nov 29 05:05:06 crc kubenswrapper[4799]: I1129 05:05:06.659704 4799 scope.go:117] "RemoveContainer" containerID="1724fe9a3398c865ba5cfd6fe24e206283bd68589be6152c56813bf9cd943125" Nov 29 05:05:06 crc kubenswrapper[4799]: E1129 05:05:06.660120 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:05:06 crc kubenswrapper[4799]: I1129 05:05:06.664705 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9a088143-0a87-46e6-bdcf-29f4150d2dfb-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d52b4\" (UID: \"9a088143-0a87-46e6-bdcf-29f4150d2dfb\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d52b4" Nov 29 05:05:06 crc kubenswrapper[4799]: I1129 05:05:06.664908 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bw7xq\" (UniqueName: \"kubernetes.io/projected/9a088143-0a87-46e6-bdcf-29f4150d2dfb-kube-api-access-bw7xq\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d52b4\" (UID: \"9a088143-0a87-46e6-bdcf-29f4150d2dfb\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d52b4" Nov 29 05:05:06 crc kubenswrapper[4799]: I1129 05:05:06.665017 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9a088143-0a87-46e6-bdcf-29f4150d2dfb-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d52b4\" (UID: \"9a088143-0a87-46e6-bdcf-29f4150d2dfb\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d52b4" Nov 29 05:05:06 crc kubenswrapper[4799]: I1129 05:05:06.697518 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zhgdz"] Nov 29 05:05:06 crc kubenswrapper[4799]: I1129 05:05:06.767134 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9a088143-0a87-46e6-bdcf-29f4150d2dfb-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d52b4\" (UID: \"9a088143-0a87-46e6-bdcf-29f4150d2dfb\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d52b4" Nov 29 05:05:06 crc kubenswrapper[4799]: I1129 05:05:06.767252 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9a088143-0a87-46e6-bdcf-29f4150d2dfb-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d52b4\" (UID: \"9a088143-0a87-46e6-bdcf-29f4150d2dfb\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d52b4" Nov 29 05:05:06 crc kubenswrapper[4799]: I1129 05:05:06.767287 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bw7xq\" (UniqueName: \"kubernetes.io/projected/9a088143-0a87-46e6-bdcf-29f4150d2dfb-kube-api-access-bw7xq\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d52b4\" (UID: \"9a088143-0a87-46e6-bdcf-29f4150d2dfb\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d52b4" Nov 29 05:05:06 crc kubenswrapper[4799]: I1129 05:05:06.773761 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9a088143-0a87-46e6-bdcf-29f4150d2dfb-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d52b4\" (UID: \"9a088143-0a87-46e6-bdcf-29f4150d2dfb\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d52b4" Nov 29 05:05:06 crc kubenswrapper[4799]: I1129 05:05:06.779043 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9a088143-0a87-46e6-bdcf-29f4150d2dfb-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d52b4\" (UID: \"9a088143-0a87-46e6-bdcf-29f4150d2dfb\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d52b4" Nov 29 05:05:06 crc kubenswrapper[4799]: I1129 05:05:06.785376 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bw7xq\" (UniqueName: \"kubernetes.io/projected/9a088143-0a87-46e6-bdcf-29f4150d2dfb-kube-api-access-bw7xq\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d52b4\" (UID: \"9a088143-0a87-46e6-bdcf-29f4150d2dfb\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d52b4" Nov 29 05:05:06 crc kubenswrapper[4799]: I1129 05:05:06.950387 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d52b4" Nov 29 05:05:07 crc kubenswrapper[4799]: I1129 05:05:07.606232 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d52b4"] Nov 29 05:05:08 crc kubenswrapper[4799]: I1129 05:05:08.523988 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d52b4" event={"ID":"9a088143-0a87-46e6-bdcf-29f4150d2dfb","Type":"ContainerStarted","Data":"0d31fc49c1ae84ac9c8d88de4c3a78fcd2f8cded3cf357bf3d2106a2faa9bc03"} Nov 29 05:05:08 crc kubenswrapper[4799]: I1129 05:05:08.524417 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zhgdz" podUID="933e9784-668a-4ffd-bf47-2206eef87d98" containerName="registry-server" containerID="cri-o://0b2075a4e9ea9576d7d07484cd53b3b187804ed0a1398eaedfe3c1195df09139" gracePeriod=2 Nov 29 05:05:09 crc kubenswrapper[4799]: I1129 05:05:09.064637 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zhgdz" Nov 29 05:05:09 crc kubenswrapper[4799]: I1129 05:05:09.137145 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b97j2\" (UniqueName: \"kubernetes.io/projected/933e9784-668a-4ffd-bf47-2206eef87d98-kube-api-access-b97j2\") pod \"933e9784-668a-4ffd-bf47-2206eef87d98\" (UID: \"933e9784-668a-4ffd-bf47-2206eef87d98\") " Nov 29 05:05:09 crc kubenswrapper[4799]: I1129 05:05:09.137330 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/933e9784-668a-4ffd-bf47-2206eef87d98-utilities\") pod \"933e9784-668a-4ffd-bf47-2206eef87d98\" (UID: \"933e9784-668a-4ffd-bf47-2206eef87d98\") " Nov 29 05:05:09 crc kubenswrapper[4799]: I1129 05:05:09.137430 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/933e9784-668a-4ffd-bf47-2206eef87d98-catalog-content\") pod \"933e9784-668a-4ffd-bf47-2206eef87d98\" (UID: \"933e9784-668a-4ffd-bf47-2206eef87d98\") " Nov 29 05:05:09 crc kubenswrapper[4799]: I1129 05:05:09.138842 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/933e9784-668a-4ffd-bf47-2206eef87d98-utilities" (OuterVolumeSpecName: "utilities") pod "933e9784-668a-4ffd-bf47-2206eef87d98" (UID: "933e9784-668a-4ffd-bf47-2206eef87d98"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:05:09 crc kubenswrapper[4799]: I1129 05:05:09.144025 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/933e9784-668a-4ffd-bf47-2206eef87d98-kube-api-access-b97j2" (OuterVolumeSpecName: "kube-api-access-b97j2") pod "933e9784-668a-4ffd-bf47-2206eef87d98" (UID: "933e9784-668a-4ffd-bf47-2206eef87d98"). InnerVolumeSpecName "kube-api-access-b97j2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:05:09 crc kubenswrapper[4799]: I1129 05:05:09.159846 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/933e9784-668a-4ffd-bf47-2206eef87d98-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "933e9784-668a-4ffd-bf47-2206eef87d98" (UID: "933e9784-668a-4ffd-bf47-2206eef87d98"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:05:09 crc kubenswrapper[4799]: I1129 05:05:09.240035 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b97j2\" (UniqueName: \"kubernetes.io/projected/933e9784-668a-4ffd-bf47-2206eef87d98-kube-api-access-b97j2\") on node \"crc\" DevicePath \"\"" Nov 29 05:05:09 crc kubenswrapper[4799]: I1129 05:05:09.240095 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/933e9784-668a-4ffd-bf47-2206eef87d98-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 05:05:09 crc kubenswrapper[4799]: I1129 05:05:09.240110 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/933e9784-668a-4ffd-bf47-2206eef87d98-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 05:05:09 crc kubenswrapper[4799]: I1129 05:05:09.536389 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d52b4" event={"ID":"9a088143-0a87-46e6-bdcf-29f4150d2dfb","Type":"ContainerStarted","Data":"dc29e098e3b691ad84aa72406975fc27debc211c0618c0de40d0d278a28c363c"} Nov 29 05:05:09 crc kubenswrapper[4799]: I1129 05:05:09.541189 4799 generic.go:334] "Generic (PLEG): container finished" podID="933e9784-668a-4ffd-bf47-2206eef87d98" containerID="0b2075a4e9ea9576d7d07484cd53b3b187804ed0a1398eaedfe3c1195df09139" exitCode=0 Nov 29 05:05:09 crc kubenswrapper[4799]: I1129 05:05:09.541268 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zhgdz" Nov 29 05:05:09 crc kubenswrapper[4799]: I1129 05:05:09.541258 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zhgdz" event={"ID":"933e9784-668a-4ffd-bf47-2206eef87d98","Type":"ContainerDied","Data":"0b2075a4e9ea9576d7d07484cd53b3b187804ed0a1398eaedfe3c1195df09139"} Nov 29 05:05:09 crc kubenswrapper[4799]: I1129 05:05:09.541439 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zhgdz" event={"ID":"933e9784-668a-4ffd-bf47-2206eef87d98","Type":"ContainerDied","Data":"2bb6a7e5a3b6cb058fb74468c7150eede1b28bdd2e0d73d229b03d8800fd586b"} Nov 29 05:05:09 crc kubenswrapper[4799]: I1129 05:05:09.541480 4799 scope.go:117] "RemoveContainer" containerID="0b2075a4e9ea9576d7d07484cd53b3b187804ed0a1398eaedfe3c1195df09139" Nov 29 05:05:09 crc kubenswrapper[4799]: I1129 05:05:09.558665 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d52b4" podStartSLOduration=2.897818296 podStartE2EDuration="3.558636341s" podCreationTimestamp="2025-11-29 05:05:06 +0000 UTC" firstStartedPulling="2025-11-29 05:05:07.609575403 +0000 UTC m=+1583.252505813" lastFinishedPulling="2025-11-29 05:05:08.270393418 +0000 UTC m=+1583.913323858" observedRunningTime="2025-11-29 05:05:09.554298035 +0000 UTC m=+1585.197228455" watchObservedRunningTime="2025-11-29 05:05:09.558636341 +0000 UTC m=+1585.201566761" Nov 29 05:05:09 crc kubenswrapper[4799]: I1129 05:05:09.582358 4799 scope.go:117] "RemoveContainer" containerID="b251c02e36322b243d69c2407b6cb40b8d7d1c61d65ff9eb365b6aba84f5f1f4" Nov 29 05:05:09 crc kubenswrapper[4799]: I1129 05:05:09.585224 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zhgdz"] Nov 29 05:05:09 crc kubenswrapper[4799]: I1129 05:05:09.594013 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zhgdz"] Nov 29 05:05:09 crc kubenswrapper[4799]: I1129 05:05:09.615423 4799 scope.go:117] "RemoveContainer" containerID="45cd14d83de548e95d17e827d295454a494bf81f0c2c93cd3fc15be15803a574" Nov 29 05:05:09 crc kubenswrapper[4799]: I1129 05:05:09.673580 4799 scope.go:117] "RemoveContainer" containerID="0b2075a4e9ea9576d7d07484cd53b3b187804ed0a1398eaedfe3c1195df09139" Nov 29 05:05:09 crc kubenswrapper[4799]: E1129 05:05:09.674681 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b2075a4e9ea9576d7d07484cd53b3b187804ed0a1398eaedfe3c1195df09139\": container with ID starting with 0b2075a4e9ea9576d7d07484cd53b3b187804ed0a1398eaedfe3c1195df09139 not found: ID does not exist" containerID="0b2075a4e9ea9576d7d07484cd53b3b187804ed0a1398eaedfe3c1195df09139" Nov 29 05:05:09 crc kubenswrapper[4799]: I1129 05:05:09.674747 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b2075a4e9ea9576d7d07484cd53b3b187804ed0a1398eaedfe3c1195df09139"} err="failed to get container status \"0b2075a4e9ea9576d7d07484cd53b3b187804ed0a1398eaedfe3c1195df09139\": rpc error: code = NotFound desc = could not find container \"0b2075a4e9ea9576d7d07484cd53b3b187804ed0a1398eaedfe3c1195df09139\": container with ID starting with 0b2075a4e9ea9576d7d07484cd53b3b187804ed0a1398eaedfe3c1195df09139 not found: ID does not exist" Nov 29 05:05:09 crc kubenswrapper[4799]: I1129 05:05:09.674807 4799 scope.go:117] "RemoveContainer" containerID="b251c02e36322b243d69c2407b6cb40b8d7d1c61d65ff9eb365b6aba84f5f1f4" Nov 29 05:05:09 crc kubenswrapper[4799]: E1129 05:05:09.675297 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b251c02e36322b243d69c2407b6cb40b8d7d1c61d65ff9eb365b6aba84f5f1f4\": container with ID starting with b251c02e36322b243d69c2407b6cb40b8d7d1c61d65ff9eb365b6aba84f5f1f4 not found: ID does not exist" containerID="b251c02e36322b243d69c2407b6cb40b8d7d1c61d65ff9eb365b6aba84f5f1f4" Nov 29 05:05:09 crc kubenswrapper[4799]: I1129 05:05:09.675362 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b251c02e36322b243d69c2407b6cb40b8d7d1c61d65ff9eb365b6aba84f5f1f4"} err="failed to get container status \"b251c02e36322b243d69c2407b6cb40b8d7d1c61d65ff9eb365b6aba84f5f1f4\": rpc error: code = NotFound desc = could not find container \"b251c02e36322b243d69c2407b6cb40b8d7d1c61d65ff9eb365b6aba84f5f1f4\": container with ID starting with b251c02e36322b243d69c2407b6cb40b8d7d1c61d65ff9eb365b6aba84f5f1f4 not found: ID does not exist" Nov 29 05:05:09 crc kubenswrapper[4799]: I1129 05:05:09.675413 4799 scope.go:117] "RemoveContainer" containerID="45cd14d83de548e95d17e827d295454a494bf81f0c2c93cd3fc15be15803a574" Nov 29 05:05:09 crc kubenswrapper[4799]: E1129 05:05:09.676083 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45cd14d83de548e95d17e827d295454a494bf81f0c2c93cd3fc15be15803a574\": container with ID starting with 45cd14d83de548e95d17e827d295454a494bf81f0c2c93cd3fc15be15803a574 not found: ID does not exist" containerID="45cd14d83de548e95d17e827d295454a494bf81f0c2c93cd3fc15be15803a574" Nov 29 05:05:09 crc kubenswrapper[4799]: I1129 05:05:09.676129 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45cd14d83de548e95d17e827d295454a494bf81f0c2c93cd3fc15be15803a574"} err="failed to get container status \"45cd14d83de548e95d17e827d295454a494bf81f0c2c93cd3fc15be15803a574\": rpc error: code = NotFound desc = could not find container \"45cd14d83de548e95d17e827d295454a494bf81f0c2c93cd3fc15be15803a574\": container with ID starting with 45cd14d83de548e95d17e827d295454a494bf81f0c2c93cd3fc15be15803a574 not found: ID does not exist" Nov 29 05:05:10 crc kubenswrapper[4799]: I1129 05:05:10.682757 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="933e9784-668a-4ffd-bf47-2206eef87d98" path="/var/lib/kubelet/pods/933e9784-668a-4ffd-bf47-2206eef87d98/volumes" Nov 29 05:05:12 crc kubenswrapper[4799]: I1129 05:05:12.577985 4799 generic.go:334] "Generic (PLEG): container finished" podID="9a088143-0a87-46e6-bdcf-29f4150d2dfb" containerID="dc29e098e3b691ad84aa72406975fc27debc211c0618c0de40d0d278a28c363c" exitCode=0 Nov 29 05:05:12 crc kubenswrapper[4799]: I1129 05:05:12.578124 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d52b4" event={"ID":"9a088143-0a87-46e6-bdcf-29f4150d2dfb","Type":"ContainerDied","Data":"dc29e098e3b691ad84aa72406975fc27debc211c0618c0de40d0d278a28c363c"} Nov 29 05:05:14 crc kubenswrapper[4799]: I1129 05:05:14.144348 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d52b4" Nov 29 05:05:14 crc kubenswrapper[4799]: I1129 05:05:14.269073 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9a088143-0a87-46e6-bdcf-29f4150d2dfb-inventory\") pod \"9a088143-0a87-46e6-bdcf-29f4150d2dfb\" (UID: \"9a088143-0a87-46e6-bdcf-29f4150d2dfb\") " Nov 29 05:05:14 crc kubenswrapper[4799]: I1129 05:05:14.269232 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9a088143-0a87-46e6-bdcf-29f4150d2dfb-ssh-key\") pod \"9a088143-0a87-46e6-bdcf-29f4150d2dfb\" (UID: \"9a088143-0a87-46e6-bdcf-29f4150d2dfb\") " Nov 29 05:05:14 crc kubenswrapper[4799]: I1129 05:05:14.269270 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bw7xq\" (UniqueName: \"kubernetes.io/projected/9a088143-0a87-46e6-bdcf-29f4150d2dfb-kube-api-access-bw7xq\") pod \"9a088143-0a87-46e6-bdcf-29f4150d2dfb\" (UID: \"9a088143-0a87-46e6-bdcf-29f4150d2dfb\") " Nov 29 05:05:14 crc kubenswrapper[4799]: I1129 05:05:14.278980 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a088143-0a87-46e6-bdcf-29f4150d2dfb-kube-api-access-bw7xq" (OuterVolumeSpecName: "kube-api-access-bw7xq") pod "9a088143-0a87-46e6-bdcf-29f4150d2dfb" (UID: "9a088143-0a87-46e6-bdcf-29f4150d2dfb"). InnerVolumeSpecName "kube-api-access-bw7xq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:05:14 crc kubenswrapper[4799]: I1129 05:05:14.297706 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a088143-0a87-46e6-bdcf-29f4150d2dfb-inventory" (OuterVolumeSpecName: "inventory") pod "9a088143-0a87-46e6-bdcf-29f4150d2dfb" (UID: "9a088143-0a87-46e6-bdcf-29f4150d2dfb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:05:14 crc kubenswrapper[4799]: I1129 05:05:14.301179 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a088143-0a87-46e6-bdcf-29f4150d2dfb-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9a088143-0a87-46e6-bdcf-29f4150d2dfb" (UID: "9a088143-0a87-46e6-bdcf-29f4150d2dfb"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:05:14 crc kubenswrapper[4799]: I1129 05:05:14.371553 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9a088143-0a87-46e6-bdcf-29f4150d2dfb-inventory\") on node \"crc\" DevicePath \"\"" Nov 29 05:05:14 crc kubenswrapper[4799]: I1129 05:05:14.371606 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9a088143-0a87-46e6-bdcf-29f4150d2dfb-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 29 05:05:14 crc kubenswrapper[4799]: I1129 05:05:14.371620 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bw7xq\" (UniqueName: \"kubernetes.io/projected/9a088143-0a87-46e6-bdcf-29f4150d2dfb-kube-api-access-bw7xq\") on node \"crc\" DevicePath \"\"" Nov 29 05:05:14 crc kubenswrapper[4799]: I1129 05:05:14.603068 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d52b4" event={"ID":"9a088143-0a87-46e6-bdcf-29f4150d2dfb","Type":"ContainerDied","Data":"0d31fc49c1ae84ac9c8d88de4c3a78fcd2f8cded3cf357bf3d2106a2faa9bc03"} Nov 29 05:05:14 crc kubenswrapper[4799]: I1129 05:05:14.603122 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d31fc49c1ae84ac9c8d88de4c3a78fcd2f8cded3cf357bf3d2106a2faa9bc03" Nov 29 05:05:14 crc kubenswrapper[4799]: I1129 05:05:14.603172 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d52b4" Nov 29 05:05:14 crc kubenswrapper[4799]: I1129 05:05:14.692254 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-s87nb"] Nov 29 05:05:14 crc kubenswrapper[4799]: E1129 05:05:14.692672 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="933e9784-668a-4ffd-bf47-2206eef87d98" containerName="extract-content" Nov 29 05:05:14 crc kubenswrapper[4799]: I1129 05:05:14.692694 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="933e9784-668a-4ffd-bf47-2206eef87d98" containerName="extract-content" Nov 29 05:05:14 crc kubenswrapper[4799]: E1129 05:05:14.692727 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a088143-0a87-46e6-bdcf-29f4150d2dfb" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 29 05:05:14 crc kubenswrapper[4799]: I1129 05:05:14.692737 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a088143-0a87-46e6-bdcf-29f4150d2dfb" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 29 05:05:14 crc kubenswrapper[4799]: E1129 05:05:14.692752 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="933e9784-668a-4ffd-bf47-2206eef87d98" containerName="extract-utilities" Nov 29 05:05:14 crc kubenswrapper[4799]: I1129 05:05:14.692763 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="933e9784-668a-4ffd-bf47-2206eef87d98" containerName="extract-utilities" Nov 29 05:05:14 crc kubenswrapper[4799]: E1129 05:05:14.692802 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="933e9784-668a-4ffd-bf47-2206eef87d98" containerName="registry-server" Nov 29 05:05:14 crc kubenswrapper[4799]: I1129 05:05:14.692809 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="933e9784-668a-4ffd-bf47-2206eef87d98" containerName="registry-server" Nov 29 05:05:14 crc kubenswrapper[4799]: I1129 05:05:14.693001 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="933e9784-668a-4ffd-bf47-2206eef87d98" containerName="registry-server" Nov 29 05:05:14 crc kubenswrapper[4799]: I1129 05:05:14.693017 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a088143-0a87-46e6-bdcf-29f4150d2dfb" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 29 05:05:14 crc kubenswrapper[4799]: I1129 05:05:14.693697 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-s87nb" Nov 29 05:05:14 crc kubenswrapper[4799]: I1129 05:05:14.696487 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 29 05:05:14 crc kubenswrapper[4799]: I1129 05:05:14.696516 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 29 05:05:14 crc kubenswrapper[4799]: I1129 05:05:14.696730 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 29 05:05:14 crc kubenswrapper[4799]: I1129 05:05:14.698107 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ljwtk" Nov 29 05:05:14 crc kubenswrapper[4799]: I1129 05:05:14.716899 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-s87nb"] Nov 29 05:05:14 crc kubenswrapper[4799]: I1129 05:05:14.780348 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/175125c0-700f-4f9c-a376-42dbbab29e5f-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-s87nb\" (UID: \"175125c0-700f-4f9c-a376-42dbbab29e5f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-s87nb" Nov 29 05:05:14 crc kubenswrapper[4799]: I1129 05:05:14.780430 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfxdz\" (UniqueName: \"kubernetes.io/projected/175125c0-700f-4f9c-a376-42dbbab29e5f-kube-api-access-lfxdz\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-s87nb\" (UID: \"175125c0-700f-4f9c-a376-42dbbab29e5f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-s87nb" Nov 29 05:05:14 crc kubenswrapper[4799]: I1129 05:05:14.780573 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/175125c0-700f-4f9c-a376-42dbbab29e5f-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-s87nb\" (UID: \"175125c0-700f-4f9c-a376-42dbbab29e5f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-s87nb" Nov 29 05:05:14 crc kubenswrapper[4799]: I1129 05:05:14.882241 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/175125c0-700f-4f9c-a376-42dbbab29e5f-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-s87nb\" (UID: \"175125c0-700f-4f9c-a376-42dbbab29e5f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-s87nb" Nov 29 05:05:14 crc kubenswrapper[4799]: I1129 05:05:14.882292 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfxdz\" (UniqueName: \"kubernetes.io/projected/175125c0-700f-4f9c-a376-42dbbab29e5f-kube-api-access-lfxdz\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-s87nb\" (UID: \"175125c0-700f-4f9c-a376-42dbbab29e5f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-s87nb" Nov 29 05:05:14 crc kubenswrapper[4799]: I1129 05:05:14.882379 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/175125c0-700f-4f9c-a376-42dbbab29e5f-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-s87nb\" (UID: \"175125c0-700f-4f9c-a376-42dbbab29e5f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-s87nb" Nov 29 05:05:14 crc kubenswrapper[4799]: I1129 05:05:14.889040 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/175125c0-700f-4f9c-a376-42dbbab29e5f-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-s87nb\" (UID: \"175125c0-700f-4f9c-a376-42dbbab29e5f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-s87nb" Nov 29 05:05:14 crc kubenswrapper[4799]: I1129 05:05:14.896046 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/175125c0-700f-4f9c-a376-42dbbab29e5f-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-s87nb\" (UID: \"175125c0-700f-4f9c-a376-42dbbab29e5f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-s87nb" Nov 29 05:05:14 crc kubenswrapper[4799]: I1129 05:05:14.900415 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfxdz\" (UniqueName: \"kubernetes.io/projected/175125c0-700f-4f9c-a376-42dbbab29e5f-kube-api-access-lfxdz\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-s87nb\" (UID: \"175125c0-700f-4f9c-a376-42dbbab29e5f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-s87nb" Nov 29 05:05:15 crc kubenswrapper[4799]: I1129 05:05:15.011914 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-s87nb" Nov 29 05:05:15 crc kubenswrapper[4799]: I1129 05:05:15.426350 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-s87nb"] Nov 29 05:05:15 crc kubenswrapper[4799]: I1129 05:05:15.615110 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-s87nb" event={"ID":"175125c0-700f-4f9c-a376-42dbbab29e5f","Type":"ContainerStarted","Data":"9afff6975b44d28b8d839ff5081082b587608dbf6ed91ab50b0a732cff84144b"} Nov 29 05:05:16 crc kubenswrapper[4799]: I1129 05:05:16.629772 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-s87nb" event={"ID":"175125c0-700f-4f9c-a376-42dbbab29e5f","Type":"ContainerStarted","Data":"290fa8684c3a97e5d407ac46c7d20d54ce882e3af94db0de6373107b07f18850"} Nov 29 05:05:16 crc kubenswrapper[4799]: I1129 05:05:16.655466 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-s87nb" podStartSLOduration=2.136683104 podStartE2EDuration="2.655445482s" podCreationTimestamp="2025-11-29 05:05:14 +0000 UTC" firstStartedPulling="2025-11-29 05:05:15.436825263 +0000 UTC m=+1591.079755663" lastFinishedPulling="2025-11-29 05:05:15.955587601 +0000 UTC m=+1591.598518041" observedRunningTime="2025-11-29 05:05:16.649720642 +0000 UTC m=+1592.292651042" watchObservedRunningTime="2025-11-29 05:05:16.655445482 +0000 UTC m=+1592.298375882" Nov 29 05:05:19 crc kubenswrapper[4799]: I1129 05:05:19.662020 4799 scope.go:117] "RemoveContainer" containerID="1724fe9a3398c865ba5cfd6fe24e206283bd68589be6152c56813bf9cd943125" Nov 29 05:05:19 crc kubenswrapper[4799]: E1129 05:05:19.663255 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:05:32 crc kubenswrapper[4799]: I1129 05:05:32.660264 4799 scope.go:117] "RemoveContainer" containerID="1724fe9a3398c865ba5cfd6fe24e206283bd68589be6152c56813bf9cd943125" Nov 29 05:05:32 crc kubenswrapper[4799]: E1129 05:05:32.661736 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:05:34 crc kubenswrapper[4799]: I1129 05:05:34.067594 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-6cz7h"] Nov 29 05:05:34 crc kubenswrapper[4799]: I1129 05:05:34.085131 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-6cz7h"] Nov 29 05:05:34 crc kubenswrapper[4799]: I1129 05:05:34.679634 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a0c30f-28ae-466a-bc63-59ad7b6afd51" path="/var/lib/kubelet/pods/57a0c30f-28ae-466a-bc63-59ad7b6afd51/volumes" Nov 29 05:05:36 crc kubenswrapper[4799]: I1129 05:05:36.039779 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-jtjzp"] Nov 29 05:05:36 crc kubenswrapper[4799]: I1129 05:05:36.050274 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-jtjzp"] Nov 29 05:05:36 crc kubenswrapper[4799]: I1129 05:05:36.677655 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96493a78-8bf7-49d4-85a0-748217f66262" path="/var/lib/kubelet/pods/96493a78-8bf7-49d4-85a0-748217f66262/volumes" Nov 29 05:05:40 crc kubenswrapper[4799]: I1129 05:05:40.042048 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-9grld"] Nov 29 05:05:40 crc kubenswrapper[4799]: I1129 05:05:40.054948 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-9grld"] Nov 29 05:05:40 crc kubenswrapper[4799]: I1129 05:05:40.674221 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bcf0c49a-5656-4399-837c-152c9af16309" path="/var/lib/kubelet/pods/bcf0c49a-5656-4399-837c-152c9af16309/volumes" Nov 29 05:05:44 crc kubenswrapper[4799]: I1129 05:05:44.659591 4799 scope.go:117] "RemoveContainer" containerID="1724fe9a3398c865ba5cfd6fe24e206283bd68589be6152c56813bf9cd943125" Nov 29 05:05:44 crc kubenswrapper[4799]: E1129 05:05:44.660640 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:05:47 crc kubenswrapper[4799]: I1129 05:05:47.055876 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-gcfzn"] Nov 29 05:05:47 crc kubenswrapper[4799]: I1129 05:05:47.075662 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-gcfzn"] Nov 29 05:05:48 crc kubenswrapper[4799]: I1129 05:05:48.675226 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8370f287-fa77-4ef5-b929-01a4cf8c598c" path="/var/lib/kubelet/pods/8370f287-fa77-4ef5-b929-01a4cf8c598c/volumes" Nov 29 05:05:50 crc kubenswrapper[4799]: I1129 05:05:50.706781 4799 scope.go:117] "RemoveContainer" containerID="bab0fee75117b65e087b33a56d3dfb3ef3bf37c7c76d8e115b396b47071a0bb7" Nov 29 05:05:50 crc kubenswrapper[4799]: I1129 05:05:50.759775 4799 scope.go:117] "RemoveContainer" containerID="b523f31635771860b61f33361291c7fdc60502a4f3a198d87bc60d8298165fdc" Nov 29 05:05:50 crc kubenswrapper[4799]: I1129 05:05:50.841925 4799 scope.go:117] "RemoveContainer" containerID="bd94990dc745ca3eae2695b4842f3b3fb903f87a853ba4e52e1a83fb34911e40" Nov 29 05:05:50 crc kubenswrapper[4799]: I1129 05:05:50.884505 4799 scope.go:117] "RemoveContainer" containerID="0786a8d8316759000427db9e93b6c94a09927610f0b98a13528749ed4a57ae4c" Nov 29 05:05:50 crc kubenswrapper[4799]: I1129 05:05:50.937513 4799 scope.go:117] "RemoveContainer" containerID="a1c1ce346c940a7f505e01f058700d6541525b7df547eeb2db3011375df3404d" Nov 29 05:05:50 crc kubenswrapper[4799]: I1129 05:05:50.984933 4799 scope.go:117] "RemoveContainer" containerID="cdcd2d082a36adb9d24d954ba9b89729c3c908a3e760003812816af54bcbb49c" Nov 29 05:05:56 crc kubenswrapper[4799]: I1129 05:05:56.659135 4799 scope.go:117] "RemoveContainer" containerID="1724fe9a3398c865ba5cfd6fe24e206283bd68589be6152c56813bf9cd943125" Nov 29 05:05:56 crc kubenswrapper[4799]: E1129 05:05:56.659973 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:05:58 crc kubenswrapper[4799]: I1129 05:05:58.060265 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-btd8j"] Nov 29 05:05:58 crc kubenswrapper[4799]: I1129 05:05:58.079050 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-btd8j"] Nov 29 05:05:58 crc kubenswrapper[4799]: I1129 05:05:58.680134 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="515c81c7-10e3-4724-aafd-42431946c2c2" path="/var/lib/kubelet/pods/515c81c7-10e3-4724-aafd-42431946c2c2/volumes" Nov 29 05:06:11 crc kubenswrapper[4799]: I1129 05:06:11.660071 4799 scope.go:117] "RemoveContainer" containerID="1724fe9a3398c865ba5cfd6fe24e206283bd68589be6152c56813bf9cd943125" Nov 29 05:06:11 crc kubenswrapper[4799]: E1129 05:06:11.661520 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:06:15 crc kubenswrapper[4799]: I1129 05:06:15.420085 4799 generic.go:334] "Generic (PLEG): container finished" podID="175125c0-700f-4f9c-a376-42dbbab29e5f" containerID="290fa8684c3a97e5d407ac46c7d20d54ce882e3af94db0de6373107b07f18850" exitCode=0 Nov 29 05:06:15 crc kubenswrapper[4799]: I1129 05:06:15.420165 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-s87nb" event={"ID":"175125c0-700f-4f9c-a376-42dbbab29e5f","Type":"ContainerDied","Data":"290fa8684c3a97e5d407ac46c7d20d54ce882e3af94db0de6373107b07f18850"} Nov 29 05:06:16 crc kubenswrapper[4799]: I1129 05:06:16.960280 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-s87nb" Nov 29 05:06:17 crc kubenswrapper[4799]: I1129 05:06:17.095365 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/175125c0-700f-4f9c-a376-42dbbab29e5f-inventory\") pod \"175125c0-700f-4f9c-a376-42dbbab29e5f\" (UID: \"175125c0-700f-4f9c-a376-42dbbab29e5f\") " Nov 29 05:06:17 crc kubenswrapper[4799]: I1129 05:06:17.095460 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lfxdz\" (UniqueName: \"kubernetes.io/projected/175125c0-700f-4f9c-a376-42dbbab29e5f-kube-api-access-lfxdz\") pod \"175125c0-700f-4f9c-a376-42dbbab29e5f\" (UID: \"175125c0-700f-4f9c-a376-42dbbab29e5f\") " Nov 29 05:06:17 crc kubenswrapper[4799]: I1129 05:06:17.095544 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/175125c0-700f-4f9c-a376-42dbbab29e5f-ssh-key\") pod \"175125c0-700f-4f9c-a376-42dbbab29e5f\" (UID: \"175125c0-700f-4f9c-a376-42dbbab29e5f\") " Nov 29 05:06:17 crc kubenswrapper[4799]: I1129 05:06:17.104488 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/175125c0-700f-4f9c-a376-42dbbab29e5f-kube-api-access-lfxdz" (OuterVolumeSpecName: "kube-api-access-lfxdz") pod "175125c0-700f-4f9c-a376-42dbbab29e5f" (UID: "175125c0-700f-4f9c-a376-42dbbab29e5f"). InnerVolumeSpecName "kube-api-access-lfxdz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:06:17 crc kubenswrapper[4799]: I1129 05:06:17.129998 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/175125c0-700f-4f9c-a376-42dbbab29e5f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "175125c0-700f-4f9c-a376-42dbbab29e5f" (UID: "175125c0-700f-4f9c-a376-42dbbab29e5f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:06:17 crc kubenswrapper[4799]: I1129 05:06:17.141580 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/175125c0-700f-4f9c-a376-42dbbab29e5f-inventory" (OuterVolumeSpecName: "inventory") pod "175125c0-700f-4f9c-a376-42dbbab29e5f" (UID: "175125c0-700f-4f9c-a376-42dbbab29e5f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:06:17 crc kubenswrapper[4799]: I1129 05:06:17.198818 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/175125c0-700f-4f9c-a376-42dbbab29e5f-inventory\") on node \"crc\" DevicePath \"\"" Nov 29 05:06:17 crc kubenswrapper[4799]: I1129 05:06:17.198867 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lfxdz\" (UniqueName: \"kubernetes.io/projected/175125c0-700f-4f9c-a376-42dbbab29e5f-kube-api-access-lfxdz\") on node \"crc\" DevicePath \"\"" Nov 29 05:06:17 crc kubenswrapper[4799]: I1129 05:06:17.198882 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/175125c0-700f-4f9c-a376-42dbbab29e5f-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 29 05:06:17 crc kubenswrapper[4799]: I1129 05:06:17.448424 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-s87nb" event={"ID":"175125c0-700f-4f9c-a376-42dbbab29e5f","Type":"ContainerDied","Data":"9afff6975b44d28b8d839ff5081082b587608dbf6ed91ab50b0a732cff84144b"} Nov 29 05:06:17 crc kubenswrapper[4799]: I1129 05:06:17.448480 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9afff6975b44d28b8d839ff5081082b587608dbf6ed91ab50b0a732cff84144b" Nov 29 05:06:17 crc kubenswrapper[4799]: I1129 05:06:17.448549 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-s87nb" Nov 29 05:06:17 crc kubenswrapper[4799]: I1129 05:06:17.576662 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-lz7r7"] Nov 29 05:06:17 crc kubenswrapper[4799]: E1129 05:06:17.577125 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="175125c0-700f-4f9c-a376-42dbbab29e5f" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 29 05:06:17 crc kubenswrapper[4799]: I1129 05:06:17.577140 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="175125c0-700f-4f9c-a376-42dbbab29e5f" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 29 05:06:17 crc kubenswrapper[4799]: I1129 05:06:17.577370 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="175125c0-700f-4f9c-a376-42dbbab29e5f" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 29 05:06:17 crc kubenswrapper[4799]: I1129 05:06:17.578046 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-lz7r7" Nov 29 05:06:17 crc kubenswrapper[4799]: I1129 05:06:17.581668 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 29 05:06:17 crc kubenswrapper[4799]: I1129 05:06:17.596599 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ljwtk" Nov 29 05:06:17 crc kubenswrapper[4799]: I1129 05:06:17.597173 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 29 05:06:17 crc kubenswrapper[4799]: I1129 05:06:17.597952 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 29 05:06:17 crc kubenswrapper[4799]: I1129 05:06:17.639197 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-lz7r7"] Nov 29 05:06:17 crc kubenswrapper[4799]: I1129 05:06:17.714197 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/55ec5d5f-5f62-4313-923c-b453223a25c4-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-lz7r7\" (UID: \"55ec5d5f-5f62-4313-923c-b453223a25c4\") " pod="openstack/ssh-known-hosts-edpm-deployment-lz7r7" Nov 29 05:06:17 crc kubenswrapper[4799]: I1129 05:06:17.714323 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hklrl\" (UniqueName: \"kubernetes.io/projected/55ec5d5f-5f62-4313-923c-b453223a25c4-kube-api-access-hklrl\") pod \"ssh-known-hosts-edpm-deployment-lz7r7\" (UID: \"55ec5d5f-5f62-4313-923c-b453223a25c4\") " pod="openstack/ssh-known-hosts-edpm-deployment-lz7r7" Nov 29 05:06:17 crc kubenswrapper[4799]: I1129 05:06:17.714525 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/55ec5d5f-5f62-4313-923c-b453223a25c4-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-lz7r7\" (UID: \"55ec5d5f-5f62-4313-923c-b453223a25c4\") " pod="openstack/ssh-known-hosts-edpm-deployment-lz7r7" Nov 29 05:06:17 crc kubenswrapper[4799]: I1129 05:06:17.816856 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/55ec5d5f-5f62-4313-923c-b453223a25c4-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-lz7r7\" (UID: \"55ec5d5f-5f62-4313-923c-b453223a25c4\") " pod="openstack/ssh-known-hosts-edpm-deployment-lz7r7" Nov 29 05:06:17 crc kubenswrapper[4799]: I1129 05:06:17.817352 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hklrl\" (UniqueName: \"kubernetes.io/projected/55ec5d5f-5f62-4313-923c-b453223a25c4-kube-api-access-hklrl\") pod \"ssh-known-hosts-edpm-deployment-lz7r7\" (UID: \"55ec5d5f-5f62-4313-923c-b453223a25c4\") " pod="openstack/ssh-known-hosts-edpm-deployment-lz7r7" Nov 29 05:06:17 crc kubenswrapper[4799]: I1129 05:06:17.817389 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/55ec5d5f-5f62-4313-923c-b453223a25c4-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-lz7r7\" (UID: \"55ec5d5f-5f62-4313-923c-b453223a25c4\") " pod="openstack/ssh-known-hosts-edpm-deployment-lz7r7" Nov 29 05:06:17 crc kubenswrapper[4799]: I1129 05:06:17.824232 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/55ec5d5f-5f62-4313-923c-b453223a25c4-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-lz7r7\" (UID: \"55ec5d5f-5f62-4313-923c-b453223a25c4\") " pod="openstack/ssh-known-hosts-edpm-deployment-lz7r7" Nov 29 05:06:17 crc kubenswrapper[4799]: I1129 05:06:17.825433 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/55ec5d5f-5f62-4313-923c-b453223a25c4-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-lz7r7\" (UID: \"55ec5d5f-5f62-4313-923c-b453223a25c4\") " pod="openstack/ssh-known-hosts-edpm-deployment-lz7r7" Nov 29 05:06:17 crc kubenswrapper[4799]: I1129 05:06:17.846120 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hklrl\" (UniqueName: \"kubernetes.io/projected/55ec5d5f-5f62-4313-923c-b453223a25c4-kube-api-access-hklrl\") pod \"ssh-known-hosts-edpm-deployment-lz7r7\" (UID: \"55ec5d5f-5f62-4313-923c-b453223a25c4\") " pod="openstack/ssh-known-hosts-edpm-deployment-lz7r7" Nov 29 05:06:17 crc kubenswrapper[4799]: I1129 05:06:17.915345 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-lz7r7" Nov 29 05:06:18 crc kubenswrapper[4799]: I1129 05:06:18.523488 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-lz7r7"] Nov 29 05:06:19 crc kubenswrapper[4799]: I1129 05:06:19.471531 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-lz7r7" event={"ID":"55ec5d5f-5f62-4313-923c-b453223a25c4","Type":"ContainerStarted","Data":"958d99d4bcd68aef2dbdf24155e268148e6336018071aaef1a1efff75d49a0fc"} Nov 29 05:06:19 crc kubenswrapper[4799]: I1129 05:06:19.472043 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-lz7r7" event={"ID":"55ec5d5f-5f62-4313-923c-b453223a25c4","Type":"ContainerStarted","Data":"8a6d8298f72c6087fc395c0dcfe127a20d7b0272638257d41444d5f4261401ca"} Nov 29 05:06:19 crc kubenswrapper[4799]: I1129 05:06:19.494638 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-lz7r7" podStartSLOduration=1.951835303 podStartE2EDuration="2.494618284s" podCreationTimestamp="2025-11-29 05:06:17 +0000 UTC" firstStartedPulling="2025-11-29 05:06:18.535477918 +0000 UTC m=+1654.178408338" lastFinishedPulling="2025-11-29 05:06:19.078260919 +0000 UTC m=+1654.721191319" observedRunningTime="2025-11-29 05:06:19.492751979 +0000 UTC m=+1655.135682399" watchObservedRunningTime="2025-11-29 05:06:19.494618284 +0000 UTC m=+1655.137548684" Nov 29 05:06:23 crc kubenswrapper[4799]: I1129 05:06:23.659562 4799 scope.go:117] "RemoveContainer" containerID="1724fe9a3398c865ba5cfd6fe24e206283bd68589be6152c56813bf9cd943125" Nov 29 05:06:23 crc kubenswrapper[4799]: E1129 05:06:23.660760 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:06:26 crc kubenswrapper[4799]: I1129 05:06:26.581233 4799 generic.go:334] "Generic (PLEG): container finished" podID="55ec5d5f-5f62-4313-923c-b453223a25c4" containerID="958d99d4bcd68aef2dbdf24155e268148e6336018071aaef1a1efff75d49a0fc" exitCode=0 Nov 29 05:06:26 crc kubenswrapper[4799]: I1129 05:06:26.581367 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-lz7r7" event={"ID":"55ec5d5f-5f62-4313-923c-b453223a25c4","Type":"ContainerDied","Data":"958d99d4bcd68aef2dbdf24155e268148e6336018071aaef1a1efff75d49a0fc"} Nov 29 05:06:28 crc kubenswrapper[4799]: I1129 05:06:28.121932 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-lz7r7" Nov 29 05:06:28 crc kubenswrapper[4799]: I1129 05:06:28.294579 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hklrl\" (UniqueName: \"kubernetes.io/projected/55ec5d5f-5f62-4313-923c-b453223a25c4-kube-api-access-hklrl\") pod \"55ec5d5f-5f62-4313-923c-b453223a25c4\" (UID: \"55ec5d5f-5f62-4313-923c-b453223a25c4\") " Nov 29 05:06:28 crc kubenswrapper[4799]: I1129 05:06:28.294891 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/55ec5d5f-5f62-4313-923c-b453223a25c4-inventory-0\") pod \"55ec5d5f-5f62-4313-923c-b453223a25c4\" (UID: \"55ec5d5f-5f62-4313-923c-b453223a25c4\") " Nov 29 05:06:28 crc kubenswrapper[4799]: I1129 05:06:28.295116 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/55ec5d5f-5f62-4313-923c-b453223a25c4-ssh-key-openstack-edpm-ipam\") pod \"55ec5d5f-5f62-4313-923c-b453223a25c4\" (UID: \"55ec5d5f-5f62-4313-923c-b453223a25c4\") " Nov 29 05:06:28 crc kubenswrapper[4799]: I1129 05:06:28.303276 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55ec5d5f-5f62-4313-923c-b453223a25c4-kube-api-access-hklrl" (OuterVolumeSpecName: "kube-api-access-hklrl") pod "55ec5d5f-5f62-4313-923c-b453223a25c4" (UID: "55ec5d5f-5f62-4313-923c-b453223a25c4"). InnerVolumeSpecName "kube-api-access-hklrl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:06:28 crc kubenswrapper[4799]: I1129 05:06:28.331786 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55ec5d5f-5f62-4313-923c-b453223a25c4-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "55ec5d5f-5f62-4313-923c-b453223a25c4" (UID: "55ec5d5f-5f62-4313-923c-b453223a25c4"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:06:28 crc kubenswrapper[4799]: I1129 05:06:28.337047 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55ec5d5f-5f62-4313-923c-b453223a25c4-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "55ec5d5f-5f62-4313-923c-b453223a25c4" (UID: "55ec5d5f-5f62-4313-923c-b453223a25c4"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:06:28 crc kubenswrapper[4799]: I1129 05:06:28.397321 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hklrl\" (UniqueName: \"kubernetes.io/projected/55ec5d5f-5f62-4313-923c-b453223a25c4-kube-api-access-hklrl\") on node \"crc\" DevicePath \"\"" Nov 29 05:06:28 crc kubenswrapper[4799]: I1129 05:06:28.397367 4799 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/55ec5d5f-5f62-4313-923c-b453223a25c4-inventory-0\") on node \"crc\" DevicePath \"\"" Nov 29 05:06:28 crc kubenswrapper[4799]: I1129 05:06:28.397388 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/55ec5d5f-5f62-4313-923c-b453223a25c4-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 29 05:06:28 crc kubenswrapper[4799]: I1129 05:06:28.614567 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-lz7r7" event={"ID":"55ec5d5f-5f62-4313-923c-b453223a25c4","Type":"ContainerDied","Data":"8a6d8298f72c6087fc395c0dcfe127a20d7b0272638257d41444d5f4261401ca"} Nov 29 05:06:28 crc kubenswrapper[4799]: I1129 05:06:28.614618 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8a6d8298f72c6087fc395c0dcfe127a20d7b0272638257d41444d5f4261401ca" Nov 29 05:06:28 crc kubenswrapper[4799]: I1129 05:06:28.614671 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-lz7r7" Nov 29 05:06:28 crc kubenswrapper[4799]: I1129 05:06:28.736834 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-cvhzj"] Nov 29 05:06:28 crc kubenswrapper[4799]: E1129 05:06:28.737825 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55ec5d5f-5f62-4313-923c-b453223a25c4" containerName="ssh-known-hosts-edpm-deployment" Nov 29 05:06:28 crc kubenswrapper[4799]: I1129 05:06:28.737847 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="55ec5d5f-5f62-4313-923c-b453223a25c4" containerName="ssh-known-hosts-edpm-deployment" Nov 29 05:06:28 crc kubenswrapper[4799]: I1129 05:06:28.738044 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="55ec5d5f-5f62-4313-923c-b453223a25c4" containerName="ssh-known-hosts-edpm-deployment" Nov 29 05:06:28 crc kubenswrapper[4799]: I1129 05:06:28.738762 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-cvhzj" Nov 29 05:06:28 crc kubenswrapper[4799]: I1129 05:06:28.746460 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 29 05:06:28 crc kubenswrapper[4799]: I1129 05:06:28.746555 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 29 05:06:28 crc kubenswrapper[4799]: I1129 05:06:28.746823 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ljwtk" Nov 29 05:06:28 crc kubenswrapper[4799]: I1129 05:06:28.746921 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 29 05:06:28 crc kubenswrapper[4799]: I1129 05:06:28.752341 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-cvhzj"] Nov 29 05:06:28 crc kubenswrapper[4799]: I1129 05:06:28.907267 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3fc9b447-c2eb-46c2-b7db-15ed96f023e3-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-cvhzj\" (UID: \"3fc9b447-c2eb-46c2-b7db-15ed96f023e3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-cvhzj" Nov 29 05:06:28 crc kubenswrapper[4799]: I1129 05:06:28.907640 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3fc9b447-c2eb-46c2-b7db-15ed96f023e3-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-cvhzj\" (UID: \"3fc9b447-c2eb-46c2-b7db-15ed96f023e3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-cvhzj" Nov 29 05:06:28 crc kubenswrapper[4799]: I1129 05:06:28.908044 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbzvz\" (UniqueName: \"kubernetes.io/projected/3fc9b447-c2eb-46c2-b7db-15ed96f023e3-kube-api-access-mbzvz\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-cvhzj\" (UID: \"3fc9b447-c2eb-46c2-b7db-15ed96f023e3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-cvhzj" Nov 29 05:06:29 crc kubenswrapper[4799]: I1129 05:06:29.010432 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3fc9b447-c2eb-46c2-b7db-15ed96f023e3-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-cvhzj\" (UID: \"3fc9b447-c2eb-46c2-b7db-15ed96f023e3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-cvhzj" Nov 29 05:06:29 crc kubenswrapper[4799]: I1129 05:06:29.010560 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbzvz\" (UniqueName: \"kubernetes.io/projected/3fc9b447-c2eb-46c2-b7db-15ed96f023e3-kube-api-access-mbzvz\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-cvhzj\" (UID: \"3fc9b447-c2eb-46c2-b7db-15ed96f023e3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-cvhzj" Nov 29 05:06:29 crc kubenswrapper[4799]: I1129 05:06:29.010696 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3fc9b447-c2eb-46c2-b7db-15ed96f023e3-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-cvhzj\" (UID: \"3fc9b447-c2eb-46c2-b7db-15ed96f023e3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-cvhzj" Nov 29 05:06:29 crc kubenswrapper[4799]: I1129 05:06:29.015520 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3fc9b447-c2eb-46c2-b7db-15ed96f023e3-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-cvhzj\" (UID: \"3fc9b447-c2eb-46c2-b7db-15ed96f023e3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-cvhzj" Nov 29 05:06:29 crc kubenswrapper[4799]: I1129 05:06:29.016851 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3fc9b447-c2eb-46c2-b7db-15ed96f023e3-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-cvhzj\" (UID: \"3fc9b447-c2eb-46c2-b7db-15ed96f023e3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-cvhzj" Nov 29 05:06:29 crc kubenswrapper[4799]: I1129 05:06:29.027469 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbzvz\" (UniqueName: \"kubernetes.io/projected/3fc9b447-c2eb-46c2-b7db-15ed96f023e3-kube-api-access-mbzvz\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-cvhzj\" (UID: \"3fc9b447-c2eb-46c2-b7db-15ed96f023e3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-cvhzj" Nov 29 05:06:29 crc kubenswrapper[4799]: I1129 05:06:29.064034 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-cvhzj" Nov 29 05:06:29 crc kubenswrapper[4799]: I1129 05:06:29.608344 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-cvhzj"] Nov 29 05:06:29 crc kubenswrapper[4799]: W1129 05:06:29.613256 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3fc9b447_c2eb_46c2_b7db_15ed96f023e3.slice/crio-079c7e9ffde61a60c17484f70e1c532868cb3d967ce00e756521ad5c1ba6337e WatchSource:0}: Error finding container 079c7e9ffde61a60c17484f70e1c532868cb3d967ce00e756521ad5c1ba6337e: Status 404 returned error can't find the container with id 079c7e9ffde61a60c17484f70e1c532868cb3d967ce00e756521ad5c1ba6337e Nov 29 05:06:29 crc kubenswrapper[4799]: I1129 05:06:29.624739 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-cvhzj" event={"ID":"3fc9b447-c2eb-46c2-b7db-15ed96f023e3","Type":"ContainerStarted","Data":"079c7e9ffde61a60c17484f70e1c532868cb3d967ce00e756521ad5c1ba6337e"} Nov 29 05:06:30 crc kubenswrapper[4799]: I1129 05:06:30.054281 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-48c8-account-create-update-t6p84"] Nov 29 05:06:30 crc kubenswrapper[4799]: I1129 05:06:30.078326 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-da5a-account-create-update-ntzr7"] Nov 29 05:06:30 crc kubenswrapper[4799]: I1129 05:06:30.087950 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-wq84l"] Nov 29 05:06:30 crc kubenswrapper[4799]: I1129 05:06:30.097866 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-hn8xp"] Nov 29 05:06:30 crc kubenswrapper[4799]: I1129 05:06:30.106804 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-6872-account-create-update-ndcpk"] Nov 29 05:06:30 crc kubenswrapper[4799]: I1129 05:06:30.125980 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-48c8-account-create-update-t6p84"] Nov 29 05:06:30 crc kubenswrapper[4799]: I1129 05:06:30.135908 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-da5a-account-create-update-ntzr7"] Nov 29 05:06:30 crc kubenswrapper[4799]: I1129 05:06:30.169501 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-zcnn2"] Nov 29 05:06:30 crc kubenswrapper[4799]: I1129 05:06:30.191955 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-wq84l"] Nov 29 05:06:30 crc kubenswrapper[4799]: I1129 05:06:30.201676 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-zcnn2"] Nov 29 05:06:30 crc kubenswrapper[4799]: I1129 05:06:30.212012 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-hn8xp"] Nov 29 05:06:30 crc kubenswrapper[4799]: I1129 05:06:30.224160 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-6872-account-create-update-ndcpk"] Nov 29 05:06:30 crc kubenswrapper[4799]: I1129 05:06:30.640070 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-cvhzj" event={"ID":"3fc9b447-c2eb-46c2-b7db-15ed96f023e3","Type":"ContainerStarted","Data":"b05742a4db8bb4e6fb8111053f044f02424c2f28699263473cb78f8841eaa088"} Nov 29 05:06:30 crc kubenswrapper[4799]: I1129 05:06:30.675114 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-cvhzj" podStartSLOduration=2.202034099 podStartE2EDuration="2.675097478s" podCreationTimestamp="2025-11-29 05:06:28 +0000 UTC" firstStartedPulling="2025-11-29 05:06:29.61912993 +0000 UTC m=+1665.262060330" lastFinishedPulling="2025-11-29 05:06:30.092193299 +0000 UTC m=+1665.735123709" observedRunningTime="2025-11-29 05:06:30.66904097 +0000 UTC m=+1666.311971370" watchObservedRunningTime="2025-11-29 05:06:30.675097478 +0000 UTC m=+1666.318027878" Nov 29 05:06:30 crc kubenswrapper[4799]: I1129 05:06:30.693545 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3452c4a6-ce88-4908-9ebf-6b53ccfcdc2e" path="/var/lib/kubelet/pods/3452c4a6-ce88-4908-9ebf-6b53ccfcdc2e/volumes" Nov 29 05:06:30 crc kubenswrapper[4799]: I1129 05:06:30.694513 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="378ff2b6-e6b0-4cb0-a6ef-33e9ce1a7021" path="/var/lib/kubelet/pods/378ff2b6-e6b0-4cb0-a6ef-33e9ce1a7021/volumes" Nov 29 05:06:30 crc kubenswrapper[4799]: I1129 05:06:30.695443 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b670080-8ca4-4e80-a1a6-ec7a63c951b6" path="/var/lib/kubelet/pods/5b670080-8ca4-4e80-a1a6-ec7a63c951b6/volumes" Nov 29 05:06:30 crc kubenswrapper[4799]: I1129 05:06:30.696359 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b88ce1ac-1c8d-407d-8183-cdac9f354607" path="/var/lib/kubelet/pods/b88ce1ac-1c8d-407d-8183-cdac9f354607/volumes" Nov 29 05:06:30 crc kubenswrapper[4799]: I1129 05:06:30.697998 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf5b318f-e3e1-45e0-b4d5-dd7fe656676f" path="/var/lib/kubelet/pods/cf5b318f-e3e1-45e0-b4d5-dd7fe656676f/volumes" Nov 29 05:06:30 crc kubenswrapper[4799]: I1129 05:06:30.702536 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3bd96cd-ca29-4a82-8c97-0709b76cd021" path="/var/lib/kubelet/pods/d3bd96cd-ca29-4a82-8c97-0709b76cd021/volumes" Nov 29 05:06:36 crc kubenswrapper[4799]: I1129 05:06:36.669864 4799 scope.go:117] "RemoveContainer" containerID="1724fe9a3398c865ba5cfd6fe24e206283bd68589be6152c56813bf9cd943125" Nov 29 05:06:36 crc kubenswrapper[4799]: E1129 05:06:36.671258 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:06:39 crc kubenswrapper[4799]: I1129 05:06:39.753554 4799 generic.go:334] "Generic (PLEG): container finished" podID="3fc9b447-c2eb-46c2-b7db-15ed96f023e3" containerID="b05742a4db8bb4e6fb8111053f044f02424c2f28699263473cb78f8841eaa088" exitCode=0 Nov 29 05:06:39 crc kubenswrapper[4799]: I1129 05:06:39.753645 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-cvhzj" event={"ID":"3fc9b447-c2eb-46c2-b7db-15ed96f023e3","Type":"ContainerDied","Data":"b05742a4db8bb4e6fb8111053f044f02424c2f28699263473cb78f8841eaa088"} Nov 29 05:06:41 crc kubenswrapper[4799]: I1129 05:06:41.358272 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-cvhzj" Nov 29 05:06:41 crc kubenswrapper[4799]: I1129 05:06:41.449503 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mbzvz\" (UniqueName: \"kubernetes.io/projected/3fc9b447-c2eb-46c2-b7db-15ed96f023e3-kube-api-access-mbzvz\") pod \"3fc9b447-c2eb-46c2-b7db-15ed96f023e3\" (UID: \"3fc9b447-c2eb-46c2-b7db-15ed96f023e3\") " Nov 29 05:06:41 crc kubenswrapper[4799]: I1129 05:06:41.449650 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3fc9b447-c2eb-46c2-b7db-15ed96f023e3-ssh-key\") pod \"3fc9b447-c2eb-46c2-b7db-15ed96f023e3\" (UID: \"3fc9b447-c2eb-46c2-b7db-15ed96f023e3\") " Nov 29 05:06:41 crc kubenswrapper[4799]: I1129 05:06:41.449770 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3fc9b447-c2eb-46c2-b7db-15ed96f023e3-inventory\") pod \"3fc9b447-c2eb-46c2-b7db-15ed96f023e3\" (UID: \"3fc9b447-c2eb-46c2-b7db-15ed96f023e3\") " Nov 29 05:06:41 crc kubenswrapper[4799]: I1129 05:06:41.460317 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fc9b447-c2eb-46c2-b7db-15ed96f023e3-kube-api-access-mbzvz" (OuterVolumeSpecName: "kube-api-access-mbzvz") pod "3fc9b447-c2eb-46c2-b7db-15ed96f023e3" (UID: "3fc9b447-c2eb-46c2-b7db-15ed96f023e3"). InnerVolumeSpecName "kube-api-access-mbzvz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:06:41 crc kubenswrapper[4799]: I1129 05:06:41.487495 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3fc9b447-c2eb-46c2-b7db-15ed96f023e3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3fc9b447-c2eb-46c2-b7db-15ed96f023e3" (UID: "3fc9b447-c2eb-46c2-b7db-15ed96f023e3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:06:41 crc kubenswrapper[4799]: I1129 05:06:41.490880 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3fc9b447-c2eb-46c2-b7db-15ed96f023e3-inventory" (OuterVolumeSpecName: "inventory") pod "3fc9b447-c2eb-46c2-b7db-15ed96f023e3" (UID: "3fc9b447-c2eb-46c2-b7db-15ed96f023e3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:06:41 crc kubenswrapper[4799]: I1129 05:06:41.552977 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mbzvz\" (UniqueName: \"kubernetes.io/projected/3fc9b447-c2eb-46c2-b7db-15ed96f023e3-kube-api-access-mbzvz\") on node \"crc\" DevicePath \"\"" Nov 29 05:06:41 crc kubenswrapper[4799]: I1129 05:06:41.553052 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3fc9b447-c2eb-46c2-b7db-15ed96f023e3-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 29 05:06:41 crc kubenswrapper[4799]: I1129 05:06:41.553069 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3fc9b447-c2eb-46c2-b7db-15ed96f023e3-inventory\") on node \"crc\" DevicePath \"\"" Nov 29 05:06:41 crc kubenswrapper[4799]: I1129 05:06:41.778667 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-cvhzj" event={"ID":"3fc9b447-c2eb-46c2-b7db-15ed96f023e3","Type":"ContainerDied","Data":"079c7e9ffde61a60c17484f70e1c532868cb3d967ce00e756521ad5c1ba6337e"} Nov 29 05:06:41 crc kubenswrapper[4799]: I1129 05:06:41.778716 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-cvhzj" Nov 29 05:06:41 crc kubenswrapper[4799]: I1129 05:06:41.778724 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="079c7e9ffde61a60c17484f70e1c532868cb3d967ce00e756521ad5c1ba6337e" Nov 29 05:06:41 crc kubenswrapper[4799]: I1129 05:06:41.888479 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qzthn"] Nov 29 05:06:41 crc kubenswrapper[4799]: E1129 05:06:41.888953 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fc9b447-c2eb-46c2-b7db-15ed96f023e3" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 29 05:06:41 crc kubenswrapper[4799]: I1129 05:06:41.888971 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fc9b447-c2eb-46c2-b7db-15ed96f023e3" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 29 05:06:41 crc kubenswrapper[4799]: I1129 05:06:41.889200 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fc9b447-c2eb-46c2-b7db-15ed96f023e3" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 29 05:06:41 crc kubenswrapper[4799]: I1129 05:06:41.890189 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qzthn" Nov 29 05:06:41 crc kubenswrapper[4799]: I1129 05:06:41.892856 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 29 05:06:41 crc kubenswrapper[4799]: I1129 05:06:41.893136 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 29 05:06:41 crc kubenswrapper[4799]: I1129 05:06:41.894816 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 29 05:06:41 crc kubenswrapper[4799]: I1129 05:06:41.894866 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ljwtk" Nov 29 05:06:41 crc kubenswrapper[4799]: I1129 05:06:41.923216 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qzthn"] Nov 29 05:06:42 crc kubenswrapper[4799]: I1129 05:06:42.066123 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d6854970-24cd-49b8-bee1-b1d3d63eeef7-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qzthn\" (UID: \"d6854970-24cd-49b8-bee1-b1d3d63eeef7\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qzthn" Nov 29 05:06:42 crc kubenswrapper[4799]: I1129 05:06:42.066746 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z76g9\" (UniqueName: \"kubernetes.io/projected/d6854970-24cd-49b8-bee1-b1d3d63eeef7-kube-api-access-z76g9\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qzthn\" (UID: \"d6854970-24cd-49b8-bee1-b1d3d63eeef7\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qzthn" Nov 29 05:06:42 crc kubenswrapper[4799]: I1129 05:06:42.066832 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d6854970-24cd-49b8-bee1-b1d3d63eeef7-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qzthn\" (UID: \"d6854970-24cd-49b8-bee1-b1d3d63eeef7\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qzthn" Nov 29 05:06:42 crc kubenswrapper[4799]: I1129 05:06:42.169312 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z76g9\" (UniqueName: \"kubernetes.io/projected/d6854970-24cd-49b8-bee1-b1d3d63eeef7-kube-api-access-z76g9\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qzthn\" (UID: \"d6854970-24cd-49b8-bee1-b1d3d63eeef7\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qzthn" Nov 29 05:06:42 crc kubenswrapper[4799]: I1129 05:06:42.169708 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d6854970-24cd-49b8-bee1-b1d3d63eeef7-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qzthn\" (UID: \"d6854970-24cd-49b8-bee1-b1d3d63eeef7\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qzthn" Nov 29 05:06:42 crc kubenswrapper[4799]: I1129 05:06:42.169821 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d6854970-24cd-49b8-bee1-b1d3d63eeef7-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qzthn\" (UID: \"d6854970-24cd-49b8-bee1-b1d3d63eeef7\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qzthn" Nov 29 05:06:42 crc kubenswrapper[4799]: I1129 05:06:42.174996 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d6854970-24cd-49b8-bee1-b1d3d63eeef7-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qzthn\" (UID: \"d6854970-24cd-49b8-bee1-b1d3d63eeef7\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qzthn" Nov 29 05:06:42 crc kubenswrapper[4799]: I1129 05:06:42.175590 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d6854970-24cd-49b8-bee1-b1d3d63eeef7-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qzthn\" (UID: \"d6854970-24cd-49b8-bee1-b1d3d63eeef7\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qzthn" Nov 29 05:06:42 crc kubenswrapper[4799]: I1129 05:06:42.196843 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z76g9\" (UniqueName: \"kubernetes.io/projected/d6854970-24cd-49b8-bee1-b1d3d63eeef7-kube-api-access-z76g9\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qzthn\" (UID: \"d6854970-24cd-49b8-bee1-b1d3d63eeef7\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qzthn" Nov 29 05:06:42 crc kubenswrapper[4799]: I1129 05:06:42.224809 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qzthn" Nov 29 05:06:43 crc kubenswrapper[4799]: I1129 05:06:42.759166 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qzthn"] Nov 29 05:06:43 crc kubenswrapper[4799]: I1129 05:06:42.826729 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qzthn" event={"ID":"d6854970-24cd-49b8-bee1-b1d3d63eeef7","Type":"ContainerStarted","Data":"de87f04a24e5720b3375c38faa32777ba6f2c7d68d5cef8aa5456dbce659944c"} Nov 29 05:06:43 crc kubenswrapper[4799]: I1129 05:06:43.841684 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qzthn" event={"ID":"d6854970-24cd-49b8-bee1-b1d3d63eeef7","Type":"ContainerStarted","Data":"9ef966072f72d4aaf4a35ed33c47a4a085d7e37260504770a37fcd0cf9f67e6f"} Nov 29 05:06:43 crc kubenswrapper[4799]: I1129 05:06:43.872741 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qzthn" podStartSLOduration=2.3646134180000002 podStartE2EDuration="2.872722682s" podCreationTimestamp="2025-11-29 05:06:41 +0000 UTC" firstStartedPulling="2025-11-29 05:06:42.772921174 +0000 UTC m=+1678.415851584" lastFinishedPulling="2025-11-29 05:06:43.281030408 +0000 UTC m=+1678.923960848" observedRunningTime="2025-11-29 05:06:43.86695642 +0000 UTC m=+1679.509886820" watchObservedRunningTime="2025-11-29 05:06:43.872722682 +0000 UTC m=+1679.515653082" Nov 29 05:06:51 crc kubenswrapper[4799]: I1129 05:06:51.184862 4799 scope.go:117] "RemoveContainer" containerID="9c14caf55dc6373fc47f3c065a099730372a3a88825444471218dabaa64b18d9" Nov 29 05:06:51 crc kubenswrapper[4799]: I1129 05:06:51.222852 4799 scope.go:117] "RemoveContainer" containerID="87cab5ee995c7c6d7b9159edcebe0c1a700965ef558a9dbd8ac8a6e1ab8b8730" Nov 29 05:06:51 crc kubenswrapper[4799]: I1129 05:06:51.355019 4799 scope.go:117] "RemoveContainer" containerID="d563bee704f704153e345a733a1a537d1c957e8507f87cf5adef9476f58065a8" Nov 29 05:06:51 crc kubenswrapper[4799]: I1129 05:06:51.381514 4799 scope.go:117] "RemoveContainer" containerID="55c318edd4521e50535e4e9250780b650068d4c774473c1d6f4882bdfa1245ae" Nov 29 05:06:51 crc kubenswrapper[4799]: I1129 05:06:51.422835 4799 scope.go:117] "RemoveContainer" containerID="2dd4967ea72c729705b7523ff2b93e3133362ad86b2a98be72716ef1eb1b4ab6" Nov 29 05:06:51 crc kubenswrapper[4799]: I1129 05:06:51.469838 4799 scope.go:117] "RemoveContainer" containerID="8674cdb7fd19ef3fdbb881b1c8e30eb93da76ed230a3598ef0d109c29ebffcf8" Nov 29 05:06:51 crc kubenswrapper[4799]: I1129 05:06:51.512635 4799 scope.go:117] "RemoveContainer" containerID="a1e80e70152500fed314f16d33879352b8ab7671cbd3309602c3c9653156594b" Nov 29 05:06:51 crc kubenswrapper[4799]: I1129 05:06:51.677536 4799 scope.go:117] "RemoveContainer" containerID="1724fe9a3398c865ba5cfd6fe24e206283bd68589be6152c56813bf9cd943125" Nov 29 05:06:51 crc kubenswrapper[4799]: E1129 05:06:51.677953 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:06:53 crc kubenswrapper[4799]: I1129 05:06:53.959398 4799 generic.go:334] "Generic (PLEG): container finished" podID="d6854970-24cd-49b8-bee1-b1d3d63eeef7" containerID="9ef966072f72d4aaf4a35ed33c47a4a085d7e37260504770a37fcd0cf9f67e6f" exitCode=0 Nov 29 05:06:53 crc kubenswrapper[4799]: I1129 05:06:53.959507 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qzthn" event={"ID":"d6854970-24cd-49b8-bee1-b1d3d63eeef7","Type":"ContainerDied","Data":"9ef966072f72d4aaf4a35ed33c47a4a085d7e37260504770a37fcd0cf9f67e6f"} Nov 29 05:06:55 crc kubenswrapper[4799]: I1129 05:06:55.516579 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qzthn" Nov 29 05:06:55 crc kubenswrapper[4799]: I1129 05:06:55.623141 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d6854970-24cd-49b8-bee1-b1d3d63eeef7-inventory\") pod \"d6854970-24cd-49b8-bee1-b1d3d63eeef7\" (UID: \"d6854970-24cd-49b8-bee1-b1d3d63eeef7\") " Nov 29 05:06:55 crc kubenswrapper[4799]: I1129 05:06:55.623302 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d6854970-24cd-49b8-bee1-b1d3d63eeef7-ssh-key\") pod \"d6854970-24cd-49b8-bee1-b1d3d63eeef7\" (UID: \"d6854970-24cd-49b8-bee1-b1d3d63eeef7\") " Nov 29 05:06:55 crc kubenswrapper[4799]: I1129 05:06:55.623430 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z76g9\" (UniqueName: \"kubernetes.io/projected/d6854970-24cd-49b8-bee1-b1d3d63eeef7-kube-api-access-z76g9\") pod \"d6854970-24cd-49b8-bee1-b1d3d63eeef7\" (UID: \"d6854970-24cd-49b8-bee1-b1d3d63eeef7\") " Nov 29 05:06:55 crc kubenswrapper[4799]: I1129 05:06:55.631887 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6854970-24cd-49b8-bee1-b1d3d63eeef7-kube-api-access-z76g9" (OuterVolumeSpecName: "kube-api-access-z76g9") pod "d6854970-24cd-49b8-bee1-b1d3d63eeef7" (UID: "d6854970-24cd-49b8-bee1-b1d3d63eeef7"). InnerVolumeSpecName "kube-api-access-z76g9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:06:55 crc kubenswrapper[4799]: I1129 05:06:55.653183 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6854970-24cd-49b8-bee1-b1d3d63eeef7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d6854970-24cd-49b8-bee1-b1d3d63eeef7" (UID: "d6854970-24cd-49b8-bee1-b1d3d63eeef7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:06:55 crc kubenswrapper[4799]: I1129 05:06:55.658095 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6854970-24cd-49b8-bee1-b1d3d63eeef7-inventory" (OuterVolumeSpecName: "inventory") pod "d6854970-24cd-49b8-bee1-b1d3d63eeef7" (UID: "d6854970-24cd-49b8-bee1-b1d3d63eeef7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:06:55 crc kubenswrapper[4799]: I1129 05:06:55.727144 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z76g9\" (UniqueName: \"kubernetes.io/projected/d6854970-24cd-49b8-bee1-b1d3d63eeef7-kube-api-access-z76g9\") on node \"crc\" DevicePath \"\"" Nov 29 05:06:55 crc kubenswrapper[4799]: I1129 05:06:55.727209 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d6854970-24cd-49b8-bee1-b1d3d63eeef7-inventory\") on node \"crc\" DevicePath \"\"" Nov 29 05:06:55 crc kubenswrapper[4799]: I1129 05:06:55.727223 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d6854970-24cd-49b8-bee1-b1d3d63eeef7-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 29 05:06:55 crc kubenswrapper[4799]: I1129 05:06:55.987120 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qzthn" event={"ID":"d6854970-24cd-49b8-bee1-b1d3d63eeef7","Type":"ContainerDied","Data":"de87f04a24e5720b3375c38faa32777ba6f2c7d68d5cef8aa5456dbce659944c"} Nov 29 05:06:55 crc kubenswrapper[4799]: I1129 05:06:55.987176 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de87f04a24e5720b3375c38faa32777ba6f2c7d68d5cef8aa5456dbce659944c" Nov 29 05:06:55 crc kubenswrapper[4799]: I1129 05:06:55.987268 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qzthn" Nov 29 05:07:01 crc kubenswrapper[4799]: I1129 05:07:01.083945 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-5g6rp"] Nov 29 05:07:01 crc kubenswrapper[4799]: I1129 05:07:01.095303 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-5g6rp"] Nov 29 05:07:02 crc kubenswrapper[4799]: I1129 05:07:02.659955 4799 scope.go:117] "RemoveContainer" containerID="1724fe9a3398c865ba5cfd6fe24e206283bd68589be6152c56813bf9cd943125" Nov 29 05:07:02 crc kubenswrapper[4799]: E1129 05:07:02.660679 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:07:02 crc kubenswrapper[4799]: I1129 05:07:02.680261 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f53c1df-8c43-4a5a-94c7-7d48fedf34ac" path="/var/lib/kubelet/pods/6f53c1df-8c43-4a5a-94c7-7d48fedf34ac/volumes" Nov 29 05:07:14 crc kubenswrapper[4799]: I1129 05:07:14.664216 4799 scope.go:117] "RemoveContainer" containerID="1724fe9a3398c865ba5cfd6fe24e206283bd68589be6152c56813bf9cd943125" Nov 29 05:07:14 crc kubenswrapper[4799]: E1129 05:07:14.665089 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:07:24 crc kubenswrapper[4799]: I1129 05:07:24.068350 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-hlvdl"] Nov 29 05:07:24 crc kubenswrapper[4799]: I1129 05:07:24.080187 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-hlvdl"] Nov 29 05:07:24 crc kubenswrapper[4799]: I1129 05:07:24.704421 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a28f76a7-54e5-4434-ba25-d3d046b30d5a" path="/var/lib/kubelet/pods/a28f76a7-54e5-4434-ba25-d3d046b30d5a/volumes" Nov 29 05:07:25 crc kubenswrapper[4799]: I1129 05:07:25.028902 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-s5l5s"] Nov 29 05:07:25 crc kubenswrapper[4799]: I1129 05:07:25.038391 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-s5l5s"] Nov 29 05:07:26 crc kubenswrapper[4799]: I1129 05:07:26.675655 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="302ec8e8-d6a9-4bf2-a586-d6e95fcb232a" path="/var/lib/kubelet/pods/302ec8e8-d6a9-4bf2-a586-d6e95fcb232a/volumes" Nov 29 05:07:28 crc kubenswrapper[4799]: I1129 05:07:28.659887 4799 scope.go:117] "RemoveContainer" containerID="1724fe9a3398c865ba5cfd6fe24e206283bd68589be6152c56813bf9cd943125" Nov 29 05:07:28 crc kubenswrapper[4799]: E1129 05:07:28.660901 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:07:42 crc kubenswrapper[4799]: I1129 05:07:42.660267 4799 scope.go:117] "RemoveContainer" containerID="1724fe9a3398c865ba5cfd6fe24e206283bd68589be6152c56813bf9cd943125" Nov 29 05:07:42 crc kubenswrapper[4799]: E1129 05:07:42.661116 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:07:51 crc kubenswrapper[4799]: I1129 05:07:51.780126 4799 scope.go:117] "RemoveContainer" containerID="d2057150feb155f1980728a394f4131b9c9ebd3a0e746e7ccd911fd293524a5c" Nov 29 05:07:51 crc kubenswrapper[4799]: I1129 05:07:51.840713 4799 scope.go:117] "RemoveContainer" containerID="405d2216d4e4ae4783cbddda45b5628a8388149b26b0386cbeb616ebef54dc2d" Nov 29 05:07:51 crc kubenswrapper[4799]: I1129 05:07:51.915769 4799 scope.go:117] "RemoveContainer" containerID="9e2c281a078f1e86deb92e6cc370f3da6dbb46757f202dd36faf43f9f9471e25" Nov 29 05:07:57 crc kubenswrapper[4799]: I1129 05:07:57.659048 4799 scope.go:117] "RemoveContainer" containerID="1724fe9a3398c865ba5cfd6fe24e206283bd68589be6152c56813bf9cd943125" Nov 29 05:07:57 crc kubenswrapper[4799]: E1129 05:07:57.659814 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:07:59 crc kubenswrapper[4799]: I1129 05:07:59.704261 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hlxn5"] Nov 29 05:07:59 crc kubenswrapper[4799]: E1129 05:07:59.706119 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6854970-24cd-49b8-bee1-b1d3d63eeef7" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 29 05:07:59 crc kubenswrapper[4799]: I1129 05:07:59.706229 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6854970-24cd-49b8-bee1-b1d3d63eeef7" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 29 05:07:59 crc kubenswrapper[4799]: I1129 05:07:59.706602 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6854970-24cd-49b8-bee1-b1d3d63eeef7" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 29 05:07:59 crc kubenswrapper[4799]: I1129 05:07:59.708696 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hlxn5" Nov 29 05:07:59 crc kubenswrapper[4799]: I1129 05:07:59.720953 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hlxn5"] Nov 29 05:07:59 crc kubenswrapper[4799]: I1129 05:07:59.849570 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eca82971-544f-4f7a-bc7d-7edb18b29258-utilities\") pod \"redhat-operators-hlxn5\" (UID: \"eca82971-544f-4f7a-bc7d-7edb18b29258\") " pod="openshift-marketplace/redhat-operators-hlxn5" Nov 29 05:07:59 crc kubenswrapper[4799]: I1129 05:07:59.849714 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eca82971-544f-4f7a-bc7d-7edb18b29258-catalog-content\") pod \"redhat-operators-hlxn5\" (UID: \"eca82971-544f-4f7a-bc7d-7edb18b29258\") " pod="openshift-marketplace/redhat-operators-hlxn5" Nov 29 05:07:59 crc kubenswrapper[4799]: I1129 05:07:59.849756 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8pwg\" (UniqueName: \"kubernetes.io/projected/eca82971-544f-4f7a-bc7d-7edb18b29258-kube-api-access-d8pwg\") pod \"redhat-operators-hlxn5\" (UID: \"eca82971-544f-4f7a-bc7d-7edb18b29258\") " pod="openshift-marketplace/redhat-operators-hlxn5" Nov 29 05:07:59 crc kubenswrapper[4799]: I1129 05:07:59.952100 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eca82971-544f-4f7a-bc7d-7edb18b29258-utilities\") pod \"redhat-operators-hlxn5\" (UID: \"eca82971-544f-4f7a-bc7d-7edb18b29258\") " pod="openshift-marketplace/redhat-operators-hlxn5" Nov 29 05:07:59 crc kubenswrapper[4799]: I1129 05:07:59.952217 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eca82971-544f-4f7a-bc7d-7edb18b29258-catalog-content\") pod \"redhat-operators-hlxn5\" (UID: \"eca82971-544f-4f7a-bc7d-7edb18b29258\") " pod="openshift-marketplace/redhat-operators-hlxn5" Nov 29 05:07:59 crc kubenswrapper[4799]: I1129 05:07:59.952248 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8pwg\" (UniqueName: \"kubernetes.io/projected/eca82971-544f-4f7a-bc7d-7edb18b29258-kube-api-access-d8pwg\") pod \"redhat-operators-hlxn5\" (UID: \"eca82971-544f-4f7a-bc7d-7edb18b29258\") " pod="openshift-marketplace/redhat-operators-hlxn5" Nov 29 05:07:59 crc kubenswrapper[4799]: I1129 05:07:59.953324 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eca82971-544f-4f7a-bc7d-7edb18b29258-utilities\") pod \"redhat-operators-hlxn5\" (UID: \"eca82971-544f-4f7a-bc7d-7edb18b29258\") " pod="openshift-marketplace/redhat-operators-hlxn5" Nov 29 05:07:59 crc kubenswrapper[4799]: I1129 05:07:59.953833 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eca82971-544f-4f7a-bc7d-7edb18b29258-catalog-content\") pod \"redhat-operators-hlxn5\" (UID: \"eca82971-544f-4f7a-bc7d-7edb18b29258\") " pod="openshift-marketplace/redhat-operators-hlxn5" Nov 29 05:07:59 crc kubenswrapper[4799]: I1129 05:07:59.994024 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8pwg\" (UniqueName: \"kubernetes.io/projected/eca82971-544f-4f7a-bc7d-7edb18b29258-kube-api-access-d8pwg\") pod \"redhat-operators-hlxn5\" (UID: \"eca82971-544f-4f7a-bc7d-7edb18b29258\") " pod="openshift-marketplace/redhat-operators-hlxn5" Nov 29 05:08:00 crc kubenswrapper[4799]: I1129 05:08:00.032160 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hlxn5" Nov 29 05:08:00 crc kubenswrapper[4799]: I1129 05:08:00.575740 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hlxn5"] Nov 29 05:08:00 crc kubenswrapper[4799]: I1129 05:08:00.798165 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hlxn5" event={"ID":"eca82971-544f-4f7a-bc7d-7edb18b29258","Type":"ContainerStarted","Data":"76bb1dc4d808decd15ad0e22f0f88c643288d77826ceb2fff7092b933a865e9f"} Nov 29 05:08:00 crc kubenswrapper[4799]: I1129 05:08:00.798230 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hlxn5" event={"ID":"eca82971-544f-4f7a-bc7d-7edb18b29258","Type":"ContainerStarted","Data":"09b6a748633552b6cc55f7b5568760c3f43643bf84415316a06281af5f6c0645"} Nov 29 05:08:01 crc kubenswrapper[4799]: I1129 05:08:01.820781 4799 generic.go:334] "Generic (PLEG): container finished" podID="eca82971-544f-4f7a-bc7d-7edb18b29258" containerID="76bb1dc4d808decd15ad0e22f0f88c643288d77826ceb2fff7092b933a865e9f" exitCode=0 Nov 29 05:08:01 crc kubenswrapper[4799]: I1129 05:08:01.821101 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hlxn5" event={"ID":"eca82971-544f-4f7a-bc7d-7edb18b29258","Type":"ContainerDied","Data":"76bb1dc4d808decd15ad0e22f0f88c643288d77826ceb2fff7092b933a865e9f"} Nov 29 05:08:02 crc kubenswrapper[4799]: I1129 05:08:02.835107 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hlxn5" event={"ID":"eca82971-544f-4f7a-bc7d-7edb18b29258","Type":"ContainerStarted","Data":"88ef589f452542decbb8554516804980ab2e907eebc546f1a792e53a09cd34fa"} Nov 29 05:08:03 crc kubenswrapper[4799]: I1129 05:08:03.854378 4799 generic.go:334] "Generic (PLEG): container finished" podID="eca82971-544f-4f7a-bc7d-7edb18b29258" containerID="88ef589f452542decbb8554516804980ab2e907eebc546f1a792e53a09cd34fa" exitCode=0 Nov 29 05:08:03 crc kubenswrapper[4799]: I1129 05:08:03.854954 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hlxn5" event={"ID":"eca82971-544f-4f7a-bc7d-7edb18b29258","Type":"ContainerDied","Data":"88ef589f452542decbb8554516804980ab2e907eebc546f1a792e53a09cd34fa"} Nov 29 05:08:04 crc kubenswrapper[4799]: I1129 05:08:04.872024 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hlxn5" event={"ID":"eca82971-544f-4f7a-bc7d-7edb18b29258","Type":"ContainerStarted","Data":"dd20bdc1a7967b946935a66ee824b65aa01a9f07ffc64f6507289331697ed1b0"} Nov 29 05:08:04 crc kubenswrapper[4799]: I1129 05:08:04.895945 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hlxn5" podStartSLOduration=3.169456081 podStartE2EDuration="5.895917327s" podCreationTimestamp="2025-11-29 05:07:59 +0000 UTC" firstStartedPulling="2025-11-29 05:08:01.833982299 +0000 UTC m=+1757.476912689" lastFinishedPulling="2025-11-29 05:08:04.560443535 +0000 UTC m=+1760.203373935" observedRunningTime="2025-11-29 05:08:04.892973055 +0000 UTC m=+1760.535903485" watchObservedRunningTime="2025-11-29 05:08:04.895917327 +0000 UTC m=+1760.538847767" Nov 29 05:08:08 crc kubenswrapper[4799]: I1129 05:08:08.659716 4799 scope.go:117] "RemoveContainer" containerID="1724fe9a3398c865ba5cfd6fe24e206283bd68589be6152c56813bf9cd943125" Nov 29 05:08:08 crc kubenswrapper[4799]: E1129 05:08:08.661182 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:08:10 crc kubenswrapper[4799]: I1129 05:08:10.032569 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hlxn5" Nov 29 05:08:10 crc kubenswrapper[4799]: I1129 05:08:10.032636 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hlxn5" Nov 29 05:08:11 crc kubenswrapper[4799]: I1129 05:08:11.068427 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-m5bjj"] Nov 29 05:08:11 crc kubenswrapper[4799]: I1129 05:08:11.088414 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hlxn5" podUID="eca82971-544f-4f7a-bc7d-7edb18b29258" containerName="registry-server" probeResult="failure" output=< Nov 29 05:08:11 crc kubenswrapper[4799]: timeout: failed to connect service ":50051" within 1s Nov 29 05:08:11 crc kubenswrapper[4799]: > Nov 29 05:08:11 crc kubenswrapper[4799]: I1129 05:08:11.090189 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-m5bjj"] Nov 29 05:08:12 crc kubenswrapper[4799]: I1129 05:08:12.675905 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35dd3ae0-d97c-49ff-8b70-a394fd58af0f" path="/var/lib/kubelet/pods/35dd3ae0-d97c-49ff-8b70-a394fd58af0f/volumes" Nov 29 05:08:20 crc kubenswrapper[4799]: I1129 05:08:20.091385 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hlxn5" Nov 29 05:08:20 crc kubenswrapper[4799]: I1129 05:08:20.148714 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hlxn5" Nov 29 05:08:20 crc kubenswrapper[4799]: I1129 05:08:20.342635 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hlxn5"] Nov 29 05:08:20 crc kubenswrapper[4799]: I1129 05:08:20.659983 4799 scope.go:117] "RemoveContainer" containerID="1724fe9a3398c865ba5cfd6fe24e206283bd68589be6152c56813bf9cd943125" Nov 29 05:08:20 crc kubenswrapper[4799]: E1129 05:08:20.660464 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:08:22 crc kubenswrapper[4799]: I1129 05:08:22.090612 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hlxn5" podUID="eca82971-544f-4f7a-bc7d-7edb18b29258" containerName="registry-server" containerID="cri-o://dd20bdc1a7967b946935a66ee824b65aa01a9f07ffc64f6507289331697ed1b0" gracePeriod=2 Nov 29 05:08:22 crc kubenswrapper[4799]: I1129 05:08:22.652479 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hlxn5" Nov 29 05:08:22 crc kubenswrapper[4799]: I1129 05:08:22.779157 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d8pwg\" (UniqueName: \"kubernetes.io/projected/eca82971-544f-4f7a-bc7d-7edb18b29258-kube-api-access-d8pwg\") pod \"eca82971-544f-4f7a-bc7d-7edb18b29258\" (UID: \"eca82971-544f-4f7a-bc7d-7edb18b29258\") " Nov 29 05:08:22 crc kubenswrapper[4799]: I1129 05:08:22.779263 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eca82971-544f-4f7a-bc7d-7edb18b29258-catalog-content\") pod \"eca82971-544f-4f7a-bc7d-7edb18b29258\" (UID: \"eca82971-544f-4f7a-bc7d-7edb18b29258\") " Nov 29 05:08:22 crc kubenswrapper[4799]: I1129 05:08:22.779402 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eca82971-544f-4f7a-bc7d-7edb18b29258-utilities\") pod \"eca82971-544f-4f7a-bc7d-7edb18b29258\" (UID: \"eca82971-544f-4f7a-bc7d-7edb18b29258\") " Nov 29 05:08:22 crc kubenswrapper[4799]: I1129 05:08:22.781880 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eca82971-544f-4f7a-bc7d-7edb18b29258-utilities" (OuterVolumeSpecName: "utilities") pod "eca82971-544f-4f7a-bc7d-7edb18b29258" (UID: "eca82971-544f-4f7a-bc7d-7edb18b29258"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:08:22 crc kubenswrapper[4799]: I1129 05:08:22.789462 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eca82971-544f-4f7a-bc7d-7edb18b29258-kube-api-access-d8pwg" (OuterVolumeSpecName: "kube-api-access-d8pwg") pod "eca82971-544f-4f7a-bc7d-7edb18b29258" (UID: "eca82971-544f-4f7a-bc7d-7edb18b29258"). InnerVolumeSpecName "kube-api-access-d8pwg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:08:22 crc kubenswrapper[4799]: I1129 05:08:22.882262 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d8pwg\" (UniqueName: \"kubernetes.io/projected/eca82971-544f-4f7a-bc7d-7edb18b29258-kube-api-access-d8pwg\") on node \"crc\" DevicePath \"\"" Nov 29 05:08:22 crc kubenswrapper[4799]: I1129 05:08:22.882320 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eca82971-544f-4f7a-bc7d-7edb18b29258-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 05:08:22 crc kubenswrapper[4799]: I1129 05:08:22.910455 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eca82971-544f-4f7a-bc7d-7edb18b29258-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eca82971-544f-4f7a-bc7d-7edb18b29258" (UID: "eca82971-544f-4f7a-bc7d-7edb18b29258"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:08:22 crc kubenswrapper[4799]: I1129 05:08:22.984715 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eca82971-544f-4f7a-bc7d-7edb18b29258-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 05:08:23 crc kubenswrapper[4799]: I1129 05:08:23.103530 4799 generic.go:334] "Generic (PLEG): container finished" podID="eca82971-544f-4f7a-bc7d-7edb18b29258" containerID="dd20bdc1a7967b946935a66ee824b65aa01a9f07ffc64f6507289331697ed1b0" exitCode=0 Nov 29 05:08:23 crc kubenswrapper[4799]: I1129 05:08:23.103580 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hlxn5" event={"ID":"eca82971-544f-4f7a-bc7d-7edb18b29258","Type":"ContainerDied","Data":"dd20bdc1a7967b946935a66ee824b65aa01a9f07ffc64f6507289331697ed1b0"} Nov 29 05:08:23 crc kubenswrapper[4799]: I1129 05:08:23.103618 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hlxn5" event={"ID":"eca82971-544f-4f7a-bc7d-7edb18b29258","Type":"ContainerDied","Data":"09b6a748633552b6cc55f7b5568760c3f43643bf84415316a06281af5f6c0645"} Nov 29 05:08:23 crc kubenswrapper[4799]: I1129 05:08:23.103657 4799 scope.go:117] "RemoveContainer" containerID="dd20bdc1a7967b946935a66ee824b65aa01a9f07ffc64f6507289331697ed1b0" Nov 29 05:08:23 crc kubenswrapper[4799]: I1129 05:08:23.103586 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hlxn5" Nov 29 05:08:23 crc kubenswrapper[4799]: I1129 05:08:23.143006 4799 scope.go:117] "RemoveContainer" containerID="88ef589f452542decbb8554516804980ab2e907eebc546f1a792e53a09cd34fa" Nov 29 05:08:23 crc kubenswrapper[4799]: I1129 05:08:23.150590 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hlxn5"] Nov 29 05:08:23 crc kubenswrapper[4799]: I1129 05:08:23.169832 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hlxn5"] Nov 29 05:08:23 crc kubenswrapper[4799]: I1129 05:08:23.183931 4799 scope.go:117] "RemoveContainer" containerID="76bb1dc4d808decd15ad0e22f0f88c643288d77826ceb2fff7092b933a865e9f" Nov 29 05:08:23 crc kubenswrapper[4799]: I1129 05:08:23.225887 4799 scope.go:117] "RemoveContainer" containerID="dd20bdc1a7967b946935a66ee824b65aa01a9f07ffc64f6507289331697ed1b0" Nov 29 05:08:23 crc kubenswrapper[4799]: E1129 05:08:23.229922 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd20bdc1a7967b946935a66ee824b65aa01a9f07ffc64f6507289331697ed1b0\": container with ID starting with dd20bdc1a7967b946935a66ee824b65aa01a9f07ffc64f6507289331697ed1b0 not found: ID does not exist" containerID="dd20bdc1a7967b946935a66ee824b65aa01a9f07ffc64f6507289331697ed1b0" Nov 29 05:08:23 crc kubenswrapper[4799]: I1129 05:08:23.229969 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd20bdc1a7967b946935a66ee824b65aa01a9f07ffc64f6507289331697ed1b0"} err="failed to get container status \"dd20bdc1a7967b946935a66ee824b65aa01a9f07ffc64f6507289331697ed1b0\": rpc error: code = NotFound desc = could not find container \"dd20bdc1a7967b946935a66ee824b65aa01a9f07ffc64f6507289331697ed1b0\": container with ID starting with dd20bdc1a7967b946935a66ee824b65aa01a9f07ffc64f6507289331697ed1b0 not found: ID does not exist" Nov 29 05:08:23 crc kubenswrapper[4799]: I1129 05:08:23.229999 4799 scope.go:117] "RemoveContainer" containerID="88ef589f452542decbb8554516804980ab2e907eebc546f1a792e53a09cd34fa" Nov 29 05:08:23 crc kubenswrapper[4799]: E1129 05:08:23.230592 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88ef589f452542decbb8554516804980ab2e907eebc546f1a792e53a09cd34fa\": container with ID starting with 88ef589f452542decbb8554516804980ab2e907eebc546f1a792e53a09cd34fa not found: ID does not exist" containerID="88ef589f452542decbb8554516804980ab2e907eebc546f1a792e53a09cd34fa" Nov 29 05:08:23 crc kubenswrapper[4799]: I1129 05:08:23.230644 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88ef589f452542decbb8554516804980ab2e907eebc546f1a792e53a09cd34fa"} err="failed to get container status \"88ef589f452542decbb8554516804980ab2e907eebc546f1a792e53a09cd34fa\": rpc error: code = NotFound desc = could not find container \"88ef589f452542decbb8554516804980ab2e907eebc546f1a792e53a09cd34fa\": container with ID starting with 88ef589f452542decbb8554516804980ab2e907eebc546f1a792e53a09cd34fa not found: ID does not exist" Nov 29 05:08:23 crc kubenswrapper[4799]: I1129 05:08:23.230678 4799 scope.go:117] "RemoveContainer" containerID="76bb1dc4d808decd15ad0e22f0f88c643288d77826ceb2fff7092b933a865e9f" Nov 29 05:08:23 crc kubenswrapper[4799]: E1129 05:08:23.231041 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76bb1dc4d808decd15ad0e22f0f88c643288d77826ceb2fff7092b933a865e9f\": container with ID starting with 76bb1dc4d808decd15ad0e22f0f88c643288d77826ceb2fff7092b933a865e9f not found: ID does not exist" containerID="76bb1dc4d808decd15ad0e22f0f88c643288d77826ceb2fff7092b933a865e9f" Nov 29 05:08:23 crc kubenswrapper[4799]: I1129 05:08:23.231072 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76bb1dc4d808decd15ad0e22f0f88c643288d77826ceb2fff7092b933a865e9f"} err="failed to get container status \"76bb1dc4d808decd15ad0e22f0f88c643288d77826ceb2fff7092b933a865e9f\": rpc error: code = NotFound desc = could not find container \"76bb1dc4d808decd15ad0e22f0f88c643288d77826ceb2fff7092b933a865e9f\": container with ID starting with 76bb1dc4d808decd15ad0e22f0f88c643288d77826ceb2fff7092b933a865e9f not found: ID does not exist" Nov 29 05:08:24 crc kubenswrapper[4799]: I1129 05:08:24.671124 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eca82971-544f-4f7a-bc7d-7edb18b29258" path="/var/lib/kubelet/pods/eca82971-544f-4f7a-bc7d-7edb18b29258/volumes" Nov 29 05:08:34 crc kubenswrapper[4799]: I1129 05:08:34.667562 4799 scope.go:117] "RemoveContainer" containerID="1724fe9a3398c865ba5cfd6fe24e206283bd68589be6152c56813bf9cd943125" Nov 29 05:08:34 crc kubenswrapper[4799]: E1129 05:08:34.668640 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:08:48 crc kubenswrapper[4799]: I1129 05:08:48.659814 4799 scope.go:117] "RemoveContainer" containerID="1724fe9a3398c865ba5cfd6fe24e206283bd68589be6152c56813bf9cd943125" Nov 29 05:08:48 crc kubenswrapper[4799]: E1129 05:08:48.660746 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:08:52 crc kubenswrapper[4799]: I1129 05:08:52.055978 4799 scope.go:117] "RemoveContainer" containerID="e1ec5147edd224f00432287e8890d08d4f3a9e3e7e0790d3533c7706df9da228" Nov 29 05:09:03 crc kubenswrapper[4799]: I1129 05:09:03.659835 4799 scope.go:117] "RemoveContainer" containerID="1724fe9a3398c865ba5cfd6fe24e206283bd68589be6152c56813bf9cd943125" Nov 29 05:09:03 crc kubenswrapper[4799]: E1129 05:09:03.661598 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:09:15 crc kubenswrapper[4799]: I1129 05:09:15.659729 4799 scope.go:117] "RemoveContainer" containerID="1724fe9a3398c865ba5cfd6fe24e206283bd68589be6152c56813bf9cd943125" Nov 29 05:09:15 crc kubenswrapper[4799]: E1129 05:09:15.660950 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:09:30 crc kubenswrapper[4799]: I1129 05:09:30.659106 4799 scope.go:117] "RemoveContainer" containerID="1724fe9a3398c865ba5cfd6fe24e206283bd68589be6152c56813bf9cd943125" Nov 29 05:09:30 crc kubenswrapper[4799]: E1129 05:09:30.660329 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:09:43 crc kubenswrapper[4799]: I1129 05:09:43.661916 4799 scope.go:117] "RemoveContainer" containerID="1724fe9a3398c865ba5cfd6fe24e206283bd68589be6152c56813bf9cd943125" Nov 29 05:09:44 crc kubenswrapper[4799]: I1129 05:09:44.049757 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" event={"ID":"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8","Type":"ContainerStarted","Data":"4a35df0e0c53bb8deb879dbc720f6c155860ecc2180303a28b55f87c720cacd9"} Nov 29 05:11:36 crc kubenswrapper[4799]: I1129 05:11:36.038930 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d52b4"] Nov 29 05:11:36 crc kubenswrapper[4799]: I1129 05:11:36.056302 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mzjpx"] Nov 29 05:11:36 crc kubenswrapper[4799]: I1129 05:11:36.062705 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mzjpx"] Nov 29 05:11:36 crc kubenswrapper[4799]: I1129 05:11:36.086410 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-5p5jb"] Nov 29 05:11:36 crc kubenswrapper[4799]: I1129 05:11:36.106278 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pvn8g"] Nov 29 05:11:36 crc kubenswrapper[4799]: I1129 05:11:36.123219 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d52b4"] Nov 29 05:11:36 crc kubenswrapper[4799]: I1129 05:11:36.133089 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-s87nb"] Nov 29 05:11:36 crc kubenswrapper[4799]: I1129 05:11:36.147445 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-lz7r7"] Nov 29 05:11:36 crc kubenswrapper[4799]: I1129 05:11:36.157578 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-5p5jb"] Nov 29 05:11:36 crc kubenswrapper[4799]: I1129 05:11:36.165990 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-whnxl"] Nov 29 05:11:36 crc kubenswrapper[4799]: I1129 05:11:36.173082 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-cvhzj"] Nov 29 05:11:36 crc kubenswrapper[4799]: I1129 05:11:36.183111 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pvn8g"] Nov 29 05:11:36 crc kubenswrapper[4799]: I1129 05:11:36.189692 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wbvkf"] Nov 29 05:11:36 crc kubenswrapper[4799]: I1129 05:11:36.195495 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-s87nb"] Nov 29 05:11:36 crc kubenswrapper[4799]: I1129 05:11:36.201371 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wbvkf"] Nov 29 05:11:36 crc kubenswrapper[4799]: I1129 05:11:36.207683 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-lz7r7"] Nov 29 05:11:36 crc kubenswrapper[4799]: I1129 05:11:36.213906 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qzthn"] Nov 29 05:11:36 crc kubenswrapper[4799]: I1129 05:11:36.219472 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qzthn"] Nov 29 05:11:36 crc kubenswrapper[4799]: I1129 05:11:36.225326 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-cvhzj"] Nov 29 05:11:36 crc kubenswrapper[4799]: I1129 05:11:36.231942 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-whnxl"] Nov 29 05:11:36 crc kubenswrapper[4799]: I1129 05:11:36.677459 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="175125c0-700f-4f9c-a376-42dbbab29e5f" path="/var/lib/kubelet/pods/175125c0-700f-4f9c-a376-42dbbab29e5f/volumes" Nov 29 05:11:36 crc kubenswrapper[4799]: I1129 05:11:36.679326 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e328538-a1ef-49dd-9ecd-d83560c36e15" path="/var/lib/kubelet/pods/2e328538-a1ef-49dd-9ecd-d83560c36e15/volumes" Nov 29 05:11:36 crc kubenswrapper[4799]: I1129 05:11:36.680545 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3fc9b447-c2eb-46c2-b7db-15ed96f023e3" path="/var/lib/kubelet/pods/3fc9b447-c2eb-46c2-b7db-15ed96f023e3/volumes" Nov 29 05:11:36 crc kubenswrapper[4799]: I1129 05:11:36.681961 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43fb992c-1708-4809-9c2b-db92bab1a7e7" path="/var/lib/kubelet/pods/43fb992c-1708-4809-9c2b-db92bab1a7e7/volumes" Nov 29 05:11:36 crc kubenswrapper[4799]: I1129 05:11:36.684238 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55ec5d5f-5f62-4313-923c-b453223a25c4" path="/var/lib/kubelet/pods/55ec5d5f-5f62-4313-923c-b453223a25c4/volumes" Nov 29 05:11:36 crc kubenswrapper[4799]: I1129 05:11:36.685446 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5eef8f30-b19c-47e4-9f26-eb2f386bc089" path="/var/lib/kubelet/pods/5eef8f30-b19c-47e4-9f26-eb2f386bc089/volumes" Nov 29 05:11:36 crc kubenswrapper[4799]: I1129 05:11:36.686878 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a088143-0a87-46e6-bdcf-29f4150d2dfb" path="/var/lib/kubelet/pods/9a088143-0a87-46e6-bdcf-29f4150d2dfb/volumes" Nov 29 05:11:36 crc kubenswrapper[4799]: I1129 05:11:36.688985 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa56b0dc-b80d-44d7-8c89-0081d33bba8f" path="/var/lib/kubelet/pods/aa56b0dc-b80d-44d7-8c89-0081d33bba8f/volumes" Nov 29 05:11:36 crc kubenswrapper[4799]: I1129 05:11:36.690223 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6854970-24cd-49b8-bee1-b1d3d63eeef7" path="/var/lib/kubelet/pods/d6854970-24cd-49b8-bee1-b1d3d63eeef7/volumes" Nov 29 05:11:36 crc kubenswrapper[4799]: I1129 05:11:36.691426 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f71ef031-f30f-45aa-9d42-eb3981250587" path="/var/lib/kubelet/pods/f71ef031-f30f-45aa-9d42-eb3981250587/volumes" Nov 29 05:11:42 crc kubenswrapper[4799]: I1129 05:11:42.213151 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vj9bq"] Nov 29 05:11:42 crc kubenswrapper[4799]: E1129 05:11:42.214248 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eca82971-544f-4f7a-bc7d-7edb18b29258" containerName="extract-content" Nov 29 05:11:42 crc kubenswrapper[4799]: I1129 05:11:42.214263 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="eca82971-544f-4f7a-bc7d-7edb18b29258" containerName="extract-content" Nov 29 05:11:42 crc kubenswrapper[4799]: E1129 05:11:42.214314 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eca82971-544f-4f7a-bc7d-7edb18b29258" containerName="extract-utilities" Nov 29 05:11:42 crc kubenswrapper[4799]: I1129 05:11:42.214321 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="eca82971-544f-4f7a-bc7d-7edb18b29258" containerName="extract-utilities" Nov 29 05:11:42 crc kubenswrapper[4799]: E1129 05:11:42.214329 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eca82971-544f-4f7a-bc7d-7edb18b29258" containerName="registry-server" Nov 29 05:11:42 crc kubenswrapper[4799]: I1129 05:11:42.214337 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="eca82971-544f-4f7a-bc7d-7edb18b29258" containerName="registry-server" Nov 29 05:11:42 crc kubenswrapper[4799]: I1129 05:11:42.214502 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="eca82971-544f-4f7a-bc7d-7edb18b29258" containerName="registry-server" Nov 29 05:11:42 crc kubenswrapper[4799]: I1129 05:11:42.215195 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vj9bq" Nov 29 05:11:42 crc kubenswrapper[4799]: I1129 05:11:42.218668 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 29 05:11:42 crc kubenswrapper[4799]: I1129 05:11:42.218721 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 29 05:11:42 crc kubenswrapper[4799]: I1129 05:11:42.220822 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 29 05:11:42 crc kubenswrapper[4799]: I1129 05:11:42.220837 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 29 05:11:42 crc kubenswrapper[4799]: I1129 05:11:42.221428 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ljwtk" Nov 29 05:11:42 crc kubenswrapper[4799]: I1129 05:11:42.239074 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vj9bq"] Nov 29 05:11:42 crc kubenswrapper[4799]: I1129 05:11:42.287907 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3b36b813-5af5-4cc9-92c9-818aa2b99423-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vj9bq\" (UID: \"3b36b813-5af5-4cc9-92c9-818aa2b99423\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vj9bq" Nov 29 05:11:42 crc kubenswrapper[4799]: I1129 05:11:42.288163 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9bf4\" (UniqueName: \"kubernetes.io/projected/3b36b813-5af5-4cc9-92c9-818aa2b99423-kube-api-access-b9bf4\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vj9bq\" (UID: \"3b36b813-5af5-4cc9-92c9-818aa2b99423\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vj9bq" Nov 29 05:11:42 crc kubenswrapper[4799]: I1129 05:11:42.288257 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3b36b813-5af5-4cc9-92c9-818aa2b99423-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vj9bq\" (UID: \"3b36b813-5af5-4cc9-92c9-818aa2b99423\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vj9bq" Nov 29 05:11:42 crc kubenswrapper[4799]: I1129 05:11:42.288289 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b36b813-5af5-4cc9-92c9-818aa2b99423-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vj9bq\" (UID: \"3b36b813-5af5-4cc9-92c9-818aa2b99423\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vj9bq" Nov 29 05:11:42 crc kubenswrapper[4799]: I1129 05:11:42.288425 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3b36b813-5af5-4cc9-92c9-818aa2b99423-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vj9bq\" (UID: \"3b36b813-5af5-4cc9-92c9-818aa2b99423\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vj9bq" Nov 29 05:11:42 crc kubenswrapper[4799]: I1129 05:11:42.391126 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3b36b813-5af5-4cc9-92c9-818aa2b99423-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vj9bq\" (UID: \"3b36b813-5af5-4cc9-92c9-818aa2b99423\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vj9bq" Nov 29 05:11:42 crc kubenswrapper[4799]: I1129 05:11:42.391358 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9bf4\" (UniqueName: \"kubernetes.io/projected/3b36b813-5af5-4cc9-92c9-818aa2b99423-kube-api-access-b9bf4\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vj9bq\" (UID: \"3b36b813-5af5-4cc9-92c9-818aa2b99423\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vj9bq" Nov 29 05:11:42 crc kubenswrapper[4799]: I1129 05:11:42.391426 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3b36b813-5af5-4cc9-92c9-818aa2b99423-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vj9bq\" (UID: \"3b36b813-5af5-4cc9-92c9-818aa2b99423\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vj9bq" Nov 29 05:11:42 crc kubenswrapper[4799]: I1129 05:11:42.391484 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b36b813-5af5-4cc9-92c9-818aa2b99423-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vj9bq\" (UID: \"3b36b813-5af5-4cc9-92c9-818aa2b99423\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vj9bq" Nov 29 05:11:42 crc kubenswrapper[4799]: I1129 05:11:42.391601 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3b36b813-5af5-4cc9-92c9-818aa2b99423-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vj9bq\" (UID: \"3b36b813-5af5-4cc9-92c9-818aa2b99423\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vj9bq" Nov 29 05:11:42 crc kubenswrapper[4799]: I1129 05:11:42.399781 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3b36b813-5af5-4cc9-92c9-818aa2b99423-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vj9bq\" (UID: \"3b36b813-5af5-4cc9-92c9-818aa2b99423\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vj9bq" Nov 29 05:11:42 crc kubenswrapper[4799]: I1129 05:11:42.401660 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3b36b813-5af5-4cc9-92c9-818aa2b99423-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vj9bq\" (UID: \"3b36b813-5af5-4cc9-92c9-818aa2b99423\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vj9bq" Nov 29 05:11:42 crc kubenswrapper[4799]: I1129 05:11:42.401997 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b36b813-5af5-4cc9-92c9-818aa2b99423-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vj9bq\" (UID: \"3b36b813-5af5-4cc9-92c9-818aa2b99423\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vj9bq" Nov 29 05:11:42 crc kubenswrapper[4799]: I1129 05:11:42.415733 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3b36b813-5af5-4cc9-92c9-818aa2b99423-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vj9bq\" (UID: \"3b36b813-5af5-4cc9-92c9-818aa2b99423\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vj9bq" Nov 29 05:11:42 crc kubenswrapper[4799]: I1129 05:11:42.416704 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9bf4\" (UniqueName: \"kubernetes.io/projected/3b36b813-5af5-4cc9-92c9-818aa2b99423-kube-api-access-b9bf4\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vj9bq\" (UID: \"3b36b813-5af5-4cc9-92c9-818aa2b99423\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vj9bq" Nov 29 05:11:42 crc kubenswrapper[4799]: I1129 05:11:42.552614 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vj9bq" Nov 29 05:11:43 crc kubenswrapper[4799]: I1129 05:11:43.187598 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vj9bq"] Nov 29 05:11:43 crc kubenswrapper[4799]: I1129 05:11:43.188761 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 29 05:11:43 crc kubenswrapper[4799]: I1129 05:11:43.597529 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vj9bq" event={"ID":"3b36b813-5af5-4cc9-92c9-818aa2b99423","Type":"ContainerStarted","Data":"ac5666e7a09c111f78d304bea2863eebb7aac2e8120fc0ba4d523c76caa38615"} Nov 29 05:11:44 crc kubenswrapper[4799]: I1129 05:11:44.611528 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vj9bq" event={"ID":"3b36b813-5af5-4cc9-92c9-818aa2b99423","Type":"ContainerStarted","Data":"390e01aedb0d785d88965a358cf64f25ab2b7a7bff483809773dd0c9d43070f3"} Nov 29 05:11:44 crc kubenswrapper[4799]: I1129 05:11:44.644173 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vj9bq" podStartSLOduration=1.930368136 podStartE2EDuration="2.64414115s" podCreationTimestamp="2025-11-29 05:11:42 +0000 UTC" firstStartedPulling="2025-11-29 05:11:43.188436021 +0000 UTC m=+1978.831366431" lastFinishedPulling="2025-11-29 05:11:43.902208995 +0000 UTC m=+1979.545139445" observedRunningTime="2025-11-29 05:11:44.63834125 +0000 UTC m=+1980.281271690" watchObservedRunningTime="2025-11-29 05:11:44.64414115 +0000 UTC m=+1980.287071590" Nov 29 05:11:52 crc kubenswrapper[4799]: I1129 05:11:52.213811 4799 scope.go:117] "RemoveContainer" containerID="290fa8684c3a97e5d407ac46c7d20d54ce882e3af94db0de6373107b07f18850" Nov 29 05:11:52 crc kubenswrapper[4799]: I1129 05:11:52.294083 4799 scope.go:117] "RemoveContainer" containerID="5d2ced6ff32786bd7820e5889d3ac4340084b3f818960ca1cd8507dbe9a6b4a6" Nov 29 05:11:52 crc kubenswrapper[4799]: I1129 05:11:52.363647 4799 scope.go:117] "RemoveContainer" containerID="3032d7d0134d29195583edf79494e0e4a42f8b7784aedd8985ff8e0591f66748" Nov 29 05:11:52 crc kubenswrapper[4799]: I1129 05:11:52.418033 4799 scope.go:117] "RemoveContainer" containerID="2d082b6293a5c8c05c0f55aec36c2335bfad32aae314d76b625580bff9247b53" Nov 29 05:11:52 crc kubenswrapper[4799]: I1129 05:11:52.486250 4799 scope.go:117] "RemoveContainer" containerID="dc29e098e3b691ad84aa72406975fc27debc211c0618c0de40d0d278a28c363c" Nov 29 05:11:52 crc kubenswrapper[4799]: I1129 05:11:52.533917 4799 scope.go:117] "RemoveContainer" containerID="f7dd788a7b11409399c220143599ae99cc4f8c536ff0e788b64946836e99500d" Nov 29 05:11:52 crc kubenswrapper[4799]: I1129 05:11:52.578716 4799 scope.go:117] "RemoveContainer" containerID="3d6883d8d8a872a0fc6be069ceab4743b12e54414dcbe264bf35dda2e1182b7e" Nov 29 05:11:56 crc kubenswrapper[4799]: I1129 05:11:56.788121 4799 generic.go:334] "Generic (PLEG): container finished" podID="3b36b813-5af5-4cc9-92c9-818aa2b99423" containerID="390e01aedb0d785d88965a358cf64f25ab2b7a7bff483809773dd0c9d43070f3" exitCode=0 Nov 29 05:11:56 crc kubenswrapper[4799]: I1129 05:11:56.788210 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vj9bq" event={"ID":"3b36b813-5af5-4cc9-92c9-818aa2b99423","Type":"ContainerDied","Data":"390e01aedb0d785d88965a358cf64f25ab2b7a7bff483809773dd0c9d43070f3"} Nov 29 05:11:58 crc kubenswrapper[4799]: I1129 05:11:58.245881 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vj9bq" Nov 29 05:11:58 crc kubenswrapper[4799]: I1129 05:11:58.386726 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3b36b813-5af5-4cc9-92c9-818aa2b99423-ceph\") pod \"3b36b813-5af5-4cc9-92c9-818aa2b99423\" (UID: \"3b36b813-5af5-4cc9-92c9-818aa2b99423\") " Nov 29 05:11:58 crc kubenswrapper[4799]: I1129 05:11:58.387266 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b9bf4\" (UniqueName: \"kubernetes.io/projected/3b36b813-5af5-4cc9-92c9-818aa2b99423-kube-api-access-b9bf4\") pod \"3b36b813-5af5-4cc9-92c9-818aa2b99423\" (UID: \"3b36b813-5af5-4cc9-92c9-818aa2b99423\") " Nov 29 05:11:58 crc kubenswrapper[4799]: I1129 05:11:58.387365 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3b36b813-5af5-4cc9-92c9-818aa2b99423-inventory\") pod \"3b36b813-5af5-4cc9-92c9-818aa2b99423\" (UID: \"3b36b813-5af5-4cc9-92c9-818aa2b99423\") " Nov 29 05:11:58 crc kubenswrapper[4799]: I1129 05:11:58.387454 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3b36b813-5af5-4cc9-92c9-818aa2b99423-ssh-key\") pod \"3b36b813-5af5-4cc9-92c9-818aa2b99423\" (UID: \"3b36b813-5af5-4cc9-92c9-818aa2b99423\") " Nov 29 05:11:58 crc kubenswrapper[4799]: I1129 05:11:58.387543 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b36b813-5af5-4cc9-92c9-818aa2b99423-repo-setup-combined-ca-bundle\") pod \"3b36b813-5af5-4cc9-92c9-818aa2b99423\" (UID: \"3b36b813-5af5-4cc9-92c9-818aa2b99423\") " Nov 29 05:11:58 crc kubenswrapper[4799]: I1129 05:11:58.393738 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b36b813-5af5-4cc9-92c9-818aa2b99423-ceph" (OuterVolumeSpecName: "ceph") pod "3b36b813-5af5-4cc9-92c9-818aa2b99423" (UID: "3b36b813-5af5-4cc9-92c9-818aa2b99423"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:11:58 crc kubenswrapper[4799]: I1129 05:11:58.395025 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b36b813-5af5-4cc9-92c9-818aa2b99423-kube-api-access-b9bf4" (OuterVolumeSpecName: "kube-api-access-b9bf4") pod "3b36b813-5af5-4cc9-92c9-818aa2b99423" (UID: "3b36b813-5af5-4cc9-92c9-818aa2b99423"). InnerVolumeSpecName "kube-api-access-b9bf4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:11:58 crc kubenswrapper[4799]: I1129 05:11:58.402003 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b36b813-5af5-4cc9-92c9-818aa2b99423-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "3b36b813-5af5-4cc9-92c9-818aa2b99423" (UID: "3b36b813-5af5-4cc9-92c9-818aa2b99423"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:11:58 crc kubenswrapper[4799]: I1129 05:11:58.424157 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b36b813-5af5-4cc9-92c9-818aa2b99423-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3b36b813-5af5-4cc9-92c9-818aa2b99423" (UID: "3b36b813-5af5-4cc9-92c9-818aa2b99423"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:11:58 crc kubenswrapper[4799]: I1129 05:11:58.434656 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b36b813-5af5-4cc9-92c9-818aa2b99423-inventory" (OuterVolumeSpecName: "inventory") pod "3b36b813-5af5-4cc9-92c9-818aa2b99423" (UID: "3b36b813-5af5-4cc9-92c9-818aa2b99423"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:11:58 crc kubenswrapper[4799]: I1129 05:11:58.490568 4799 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3b36b813-5af5-4cc9-92c9-818aa2b99423-ceph\") on node \"crc\" DevicePath \"\"" Nov 29 05:11:58 crc kubenswrapper[4799]: I1129 05:11:58.490619 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b9bf4\" (UniqueName: \"kubernetes.io/projected/3b36b813-5af5-4cc9-92c9-818aa2b99423-kube-api-access-b9bf4\") on node \"crc\" DevicePath \"\"" Nov 29 05:11:58 crc kubenswrapper[4799]: I1129 05:11:58.490633 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3b36b813-5af5-4cc9-92c9-818aa2b99423-inventory\") on node \"crc\" DevicePath \"\"" Nov 29 05:11:58 crc kubenswrapper[4799]: I1129 05:11:58.490644 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3b36b813-5af5-4cc9-92c9-818aa2b99423-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 29 05:11:58 crc kubenswrapper[4799]: I1129 05:11:58.490658 4799 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b36b813-5af5-4cc9-92c9-818aa2b99423-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 05:11:58 crc kubenswrapper[4799]: I1129 05:11:58.809366 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vj9bq" event={"ID":"3b36b813-5af5-4cc9-92c9-818aa2b99423","Type":"ContainerDied","Data":"ac5666e7a09c111f78d304bea2863eebb7aac2e8120fc0ba4d523c76caa38615"} Nov 29 05:11:58 crc kubenswrapper[4799]: I1129 05:11:58.809842 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ac5666e7a09c111f78d304bea2863eebb7aac2e8120fc0ba4d523c76caa38615" Nov 29 05:11:58 crc kubenswrapper[4799]: I1129 05:11:58.809404 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vj9bq" Nov 29 05:11:58 crc kubenswrapper[4799]: I1129 05:11:58.883886 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9rvx"] Nov 29 05:11:58 crc kubenswrapper[4799]: E1129 05:11:58.884472 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b36b813-5af5-4cc9-92c9-818aa2b99423" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 29 05:11:58 crc kubenswrapper[4799]: I1129 05:11:58.884498 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b36b813-5af5-4cc9-92c9-818aa2b99423" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 29 05:11:58 crc kubenswrapper[4799]: I1129 05:11:58.884710 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b36b813-5af5-4cc9-92c9-818aa2b99423" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 29 05:11:58 crc kubenswrapper[4799]: I1129 05:11:58.885612 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9rvx" Nov 29 05:11:58 crc kubenswrapper[4799]: I1129 05:11:58.890629 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 29 05:11:58 crc kubenswrapper[4799]: I1129 05:11:58.890732 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 29 05:11:58 crc kubenswrapper[4799]: I1129 05:11:58.890672 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 29 05:11:58 crc kubenswrapper[4799]: I1129 05:11:58.892086 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 29 05:11:58 crc kubenswrapper[4799]: I1129 05:11:58.892168 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ljwtk" Nov 29 05:11:58 crc kubenswrapper[4799]: I1129 05:11:58.894667 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9rvx"] Nov 29 05:11:59 crc kubenswrapper[4799]: I1129 05:11:59.002086 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a30efa1d-c61a-4bc1-9350-fa1059e7bb71-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k9rvx\" (UID: \"a30efa1d-c61a-4bc1-9350-fa1059e7bb71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9rvx" Nov 29 05:11:59 crc kubenswrapper[4799]: I1129 05:11:59.002225 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a30efa1d-c61a-4bc1-9350-fa1059e7bb71-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k9rvx\" (UID: \"a30efa1d-c61a-4bc1-9350-fa1059e7bb71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9rvx" Nov 29 05:11:59 crc kubenswrapper[4799]: I1129 05:11:59.002262 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnfgp\" (UniqueName: \"kubernetes.io/projected/a30efa1d-c61a-4bc1-9350-fa1059e7bb71-kube-api-access-qnfgp\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k9rvx\" (UID: \"a30efa1d-c61a-4bc1-9350-fa1059e7bb71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9rvx" Nov 29 05:11:59 crc kubenswrapper[4799]: I1129 05:11:59.002296 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a30efa1d-c61a-4bc1-9350-fa1059e7bb71-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k9rvx\" (UID: \"a30efa1d-c61a-4bc1-9350-fa1059e7bb71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9rvx" Nov 29 05:11:59 crc kubenswrapper[4799]: I1129 05:11:59.002353 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a30efa1d-c61a-4bc1-9350-fa1059e7bb71-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k9rvx\" (UID: \"a30efa1d-c61a-4bc1-9350-fa1059e7bb71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9rvx" Nov 29 05:11:59 crc kubenswrapper[4799]: I1129 05:11:59.104914 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a30efa1d-c61a-4bc1-9350-fa1059e7bb71-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k9rvx\" (UID: \"a30efa1d-c61a-4bc1-9350-fa1059e7bb71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9rvx" Nov 29 05:11:59 crc kubenswrapper[4799]: I1129 05:11:59.105124 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a30efa1d-c61a-4bc1-9350-fa1059e7bb71-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k9rvx\" (UID: \"a30efa1d-c61a-4bc1-9350-fa1059e7bb71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9rvx" Nov 29 05:11:59 crc kubenswrapper[4799]: I1129 05:11:59.105175 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnfgp\" (UniqueName: \"kubernetes.io/projected/a30efa1d-c61a-4bc1-9350-fa1059e7bb71-kube-api-access-qnfgp\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k9rvx\" (UID: \"a30efa1d-c61a-4bc1-9350-fa1059e7bb71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9rvx" Nov 29 05:11:59 crc kubenswrapper[4799]: I1129 05:11:59.105223 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a30efa1d-c61a-4bc1-9350-fa1059e7bb71-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k9rvx\" (UID: \"a30efa1d-c61a-4bc1-9350-fa1059e7bb71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9rvx" Nov 29 05:11:59 crc kubenswrapper[4799]: I1129 05:11:59.105308 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a30efa1d-c61a-4bc1-9350-fa1059e7bb71-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k9rvx\" (UID: \"a30efa1d-c61a-4bc1-9350-fa1059e7bb71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9rvx" Nov 29 05:11:59 crc kubenswrapper[4799]: I1129 05:11:59.110818 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a30efa1d-c61a-4bc1-9350-fa1059e7bb71-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k9rvx\" (UID: \"a30efa1d-c61a-4bc1-9350-fa1059e7bb71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9rvx" Nov 29 05:11:59 crc kubenswrapper[4799]: I1129 05:11:59.114832 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a30efa1d-c61a-4bc1-9350-fa1059e7bb71-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k9rvx\" (UID: \"a30efa1d-c61a-4bc1-9350-fa1059e7bb71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9rvx" Nov 29 05:11:59 crc kubenswrapper[4799]: I1129 05:11:59.116014 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a30efa1d-c61a-4bc1-9350-fa1059e7bb71-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k9rvx\" (UID: \"a30efa1d-c61a-4bc1-9350-fa1059e7bb71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9rvx" Nov 29 05:11:59 crc kubenswrapper[4799]: I1129 05:11:59.124781 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a30efa1d-c61a-4bc1-9350-fa1059e7bb71-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k9rvx\" (UID: \"a30efa1d-c61a-4bc1-9350-fa1059e7bb71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9rvx" Nov 29 05:11:59 crc kubenswrapper[4799]: I1129 05:11:59.128224 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnfgp\" (UniqueName: \"kubernetes.io/projected/a30efa1d-c61a-4bc1-9350-fa1059e7bb71-kube-api-access-qnfgp\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k9rvx\" (UID: \"a30efa1d-c61a-4bc1-9350-fa1059e7bb71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9rvx" Nov 29 05:11:59 crc kubenswrapper[4799]: I1129 05:11:59.212272 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9rvx" Nov 29 05:11:59 crc kubenswrapper[4799]: I1129 05:11:59.786288 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9rvx"] Nov 29 05:11:59 crc kubenswrapper[4799]: I1129 05:11:59.820474 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9rvx" event={"ID":"a30efa1d-c61a-4bc1-9350-fa1059e7bb71","Type":"ContainerStarted","Data":"7f998d692851c3534a5c5b353947a03ff1a5c00d216f846fcb4cb70fab28a9aa"} Nov 29 05:12:00 crc kubenswrapper[4799]: I1129 05:12:00.833835 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9rvx" event={"ID":"a30efa1d-c61a-4bc1-9350-fa1059e7bb71","Type":"ContainerStarted","Data":"080bbaaf857556bcacf7aa290003173eb8109bb9d0f567010ad3c20de0a0ac78"} Nov 29 05:12:00 crc kubenswrapper[4799]: I1129 05:12:00.875647 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9rvx" podStartSLOduration=2.362683783 podStartE2EDuration="2.875617273s" podCreationTimestamp="2025-11-29 05:11:58 +0000 UTC" firstStartedPulling="2025-11-29 05:11:59.797131899 +0000 UTC m=+1995.440062319" lastFinishedPulling="2025-11-29 05:12:00.310065369 +0000 UTC m=+1995.952995809" observedRunningTime="2025-11-29 05:12:00.860263041 +0000 UTC m=+1996.503193471" watchObservedRunningTime="2025-11-29 05:12:00.875617273 +0000 UTC m=+1996.518547673" Nov 29 05:12:07 crc kubenswrapper[4799]: I1129 05:12:07.498586 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 05:12:07 crc kubenswrapper[4799]: I1129 05:12:07.501433 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 05:12:37 crc kubenswrapper[4799]: I1129 05:12:37.498069 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 05:12:37 crc kubenswrapper[4799]: I1129 05:12:37.498860 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 05:12:52 crc kubenswrapper[4799]: I1129 05:12:52.752485 4799 scope.go:117] "RemoveContainer" containerID="9ef966072f72d4aaf4a35ed33c47a4a085d7e37260504770a37fcd0cf9f67e6f" Nov 29 05:12:52 crc kubenswrapper[4799]: I1129 05:12:52.811967 4799 scope.go:117] "RemoveContainer" containerID="b05742a4db8bb4e6fb8111053f044f02424c2f28699263473cb78f8841eaa088" Nov 29 05:12:52 crc kubenswrapper[4799]: I1129 05:12:52.872003 4799 scope.go:117] "RemoveContainer" containerID="958d99d4bcd68aef2dbdf24155e268148e6336018071aaef1a1efff75d49a0fc" Nov 29 05:13:07 crc kubenswrapper[4799]: I1129 05:13:07.498370 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 05:13:07 crc kubenswrapper[4799]: I1129 05:13:07.499229 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 05:13:07 crc kubenswrapper[4799]: I1129 05:13:07.499294 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-26t88" Nov 29 05:13:07 crc kubenswrapper[4799]: I1129 05:13:07.500302 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4a35df0e0c53bb8deb879dbc720f6c155860ecc2180303a28b55f87c720cacd9"} pod="openshift-machine-config-operator/machine-config-daemon-26t88" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 29 05:13:07 crc kubenswrapper[4799]: I1129 05:13:07.500369 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" containerID="cri-o://4a35df0e0c53bb8deb879dbc720f6c155860ecc2180303a28b55f87c720cacd9" gracePeriod=600 Nov 29 05:13:08 crc kubenswrapper[4799]: I1129 05:13:08.599179 4799 generic.go:334] "Generic (PLEG): container finished" podID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerID="4a35df0e0c53bb8deb879dbc720f6c155860ecc2180303a28b55f87c720cacd9" exitCode=0 Nov 29 05:13:08 crc kubenswrapper[4799]: I1129 05:13:08.599243 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" event={"ID":"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8","Type":"ContainerDied","Data":"4a35df0e0c53bb8deb879dbc720f6c155860ecc2180303a28b55f87c720cacd9"} Nov 29 05:13:08 crc kubenswrapper[4799]: I1129 05:13:08.599722 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" event={"ID":"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8","Type":"ContainerStarted","Data":"d0281928bf1426881365580d5726280995a377a80c626b0654de38a8828878e1"} Nov 29 05:13:08 crc kubenswrapper[4799]: I1129 05:13:08.599758 4799 scope.go:117] "RemoveContainer" containerID="1724fe9a3398c865ba5cfd6fe24e206283bd68589be6152c56813bf9cd943125" Nov 29 05:13:40 crc kubenswrapper[4799]: I1129 05:13:40.961402 4799 generic.go:334] "Generic (PLEG): container finished" podID="a30efa1d-c61a-4bc1-9350-fa1059e7bb71" containerID="080bbaaf857556bcacf7aa290003173eb8109bb9d0f567010ad3c20de0a0ac78" exitCode=0 Nov 29 05:13:40 crc kubenswrapper[4799]: I1129 05:13:40.961472 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9rvx" event={"ID":"a30efa1d-c61a-4bc1-9350-fa1059e7bb71","Type":"ContainerDied","Data":"080bbaaf857556bcacf7aa290003173eb8109bb9d0f567010ad3c20de0a0ac78"} Nov 29 05:13:42 crc kubenswrapper[4799]: I1129 05:13:42.510425 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9rvx" Nov 29 05:13:42 crc kubenswrapper[4799]: I1129 05:13:42.682279 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a30efa1d-c61a-4bc1-9350-fa1059e7bb71-inventory\") pod \"a30efa1d-c61a-4bc1-9350-fa1059e7bb71\" (UID: \"a30efa1d-c61a-4bc1-9350-fa1059e7bb71\") " Nov 29 05:13:42 crc kubenswrapper[4799]: I1129 05:13:42.682386 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qnfgp\" (UniqueName: \"kubernetes.io/projected/a30efa1d-c61a-4bc1-9350-fa1059e7bb71-kube-api-access-qnfgp\") pod \"a30efa1d-c61a-4bc1-9350-fa1059e7bb71\" (UID: \"a30efa1d-c61a-4bc1-9350-fa1059e7bb71\") " Nov 29 05:13:42 crc kubenswrapper[4799]: I1129 05:13:42.682511 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a30efa1d-c61a-4bc1-9350-fa1059e7bb71-ceph\") pod \"a30efa1d-c61a-4bc1-9350-fa1059e7bb71\" (UID: \"a30efa1d-c61a-4bc1-9350-fa1059e7bb71\") " Nov 29 05:13:42 crc kubenswrapper[4799]: I1129 05:13:42.682686 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a30efa1d-c61a-4bc1-9350-fa1059e7bb71-bootstrap-combined-ca-bundle\") pod \"a30efa1d-c61a-4bc1-9350-fa1059e7bb71\" (UID: \"a30efa1d-c61a-4bc1-9350-fa1059e7bb71\") " Nov 29 05:13:42 crc kubenswrapper[4799]: I1129 05:13:42.682773 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a30efa1d-c61a-4bc1-9350-fa1059e7bb71-ssh-key\") pod \"a30efa1d-c61a-4bc1-9350-fa1059e7bb71\" (UID: \"a30efa1d-c61a-4bc1-9350-fa1059e7bb71\") " Nov 29 05:13:42 crc kubenswrapper[4799]: I1129 05:13:42.692708 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a30efa1d-c61a-4bc1-9350-fa1059e7bb71-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "a30efa1d-c61a-4bc1-9350-fa1059e7bb71" (UID: "a30efa1d-c61a-4bc1-9350-fa1059e7bb71"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:13:42 crc kubenswrapper[4799]: I1129 05:13:42.697290 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a30efa1d-c61a-4bc1-9350-fa1059e7bb71-kube-api-access-qnfgp" (OuterVolumeSpecName: "kube-api-access-qnfgp") pod "a30efa1d-c61a-4bc1-9350-fa1059e7bb71" (UID: "a30efa1d-c61a-4bc1-9350-fa1059e7bb71"). InnerVolumeSpecName "kube-api-access-qnfgp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:13:42 crc kubenswrapper[4799]: I1129 05:13:42.697291 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a30efa1d-c61a-4bc1-9350-fa1059e7bb71-ceph" (OuterVolumeSpecName: "ceph") pod "a30efa1d-c61a-4bc1-9350-fa1059e7bb71" (UID: "a30efa1d-c61a-4bc1-9350-fa1059e7bb71"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:13:42 crc kubenswrapper[4799]: I1129 05:13:42.718113 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a30efa1d-c61a-4bc1-9350-fa1059e7bb71-inventory" (OuterVolumeSpecName: "inventory") pod "a30efa1d-c61a-4bc1-9350-fa1059e7bb71" (UID: "a30efa1d-c61a-4bc1-9350-fa1059e7bb71"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:13:42 crc kubenswrapper[4799]: I1129 05:13:42.722530 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a30efa1d-c61a-4bc1-9350-fa1059e7bb71-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a30efa1d-c61a-4bc1-9350-fa1059e7bb71" (UID: "a30efa1d-c61a-4bc1-9350-fa1059e7bb71"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:13:42 crc kubenswrapper[4799]: I1129 05:13:42.786152 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a30efa1d-c61a-4bc1-9350-fa1059e7bb71-inventory\") on node \"crc\" DevicePath \"\"" Nov 29 05:13:42 crc kubenswrapper[4799]: I1129 05:13:42.786214 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qnfgp\" (UniqueName: \"kubernetes.io/projected/a30efa1d-c61a-4bc1-9350-fa1059e7bb71-kube-api-access-qnfgp\") on node \"crc\" DevicePath \"\"" Nov 29 05:13:42 crc kubenswrapper[4799]: I1129 05:13:42.786240 4799 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a30efa1d-c61a-4bc1-9350-fa1059e7bb71-ceph\") on node \"crc\" DevicePath \"\"" Nov 29 05:13:42 crc kubenswrapper[4799]: I1129 05:13:42.786260 4799 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a30efa1d-c61a-4bc1-9350-fa1059e7bb71-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 05:13:42 crc kubenswrapper[4799]: I1129 05:13:42.786278 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a30efa1d-c61a-4bc1-9350-fa1059e7bb71-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 29 05:13:42 crc kubenswrapper[4799]: I1129 05:13:42.981897 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9rvx" event={"ID":"a30efa1d-c61a-4bc1-9350-fa1059e7bb71","Type":"ContainerDied","Data":"7f998d692851c3534a5c5b353947a03ff1a5c00d216f846fcb4cb70fab28a9aa"} Nov 29 05:13:42 crc kubenswrapper[4799]: I1129 05:13:42.981946 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f998d692851c3534a5c5b353947a03ff1a5c00d216f846fcb4cb70fab28a9aa" Nov 29 05:13:42 crc kubenswrapper[4799]: I1129 05:13:42.982026 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k9rvx" Nov 29 05:13:43 crc kubenswrapper[4799]: I1129 05:13:43.147640 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x4w52"] Nov 29 05:13:43 crc kubenswrapper[4799]: E1129 05:13:43.150033 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a30efa1d-c61a-4bc1-9350-fa1059e7bb71" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 29 05:13:43 crc kubenswrapper[4799]: I1129 05:13:43.150098 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a30efa1d-c61a-4bc1-9350-fa1059e7bb71" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 29 05:13:43 crc kubenswrapper[4799]: I1129 05:13:43.151118 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="a30efa1d-c61a-4bc1-9350-fa1059e7bb71" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 29 05:13:43 crc kubenswrapper[4799]: I1129 05:13:43.153643 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x4w52" Nov 29 05:13:43 crc kubenswrapper[4799]: I1129 05:13:43.155402 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 29 05:13:43 crc kubenswrapper[4799]: I1129 05:13:43.155387 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 29 05:13:43 crc kubenswrapper[4799]: I1129 05:13:43.158242 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x4w52"] Nov 29 05:13:43 crc kubenswrapper[4799]: I1129 05:13:43.158596 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 29 05:13:43 crc kubenswrapper[4799]: I1129 05:13:43.158619 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 29 05:13:43 crc kubenswrapper[4799]: I1129 05:13:43.159283 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ljwtk" Nov 29 05:13:43 crc kubenswrapper[4799]: I1129 05:13:43.300958 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ng8tz\" (UniqueName: \"kubernetes.io/projected/86cf64fe-3c75-4b05-8503-a4e3f3e0395c-kube-api-access-ng8tz\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-x4w52\" (UID: \"86cf64fe-3c75-4b05-8503-a4e3f3e0395c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x4w52" Nov 29 05:13:43 crc kubenswrapper[4799]: I1129 05:13:43.301138 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/86cf64fe-3c75-4b05-8503-a4e3f3e0395c-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-x4w52\" (UID: \"86cf64fe-3c75-4b05-8503-a4e3f3e0395c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x4w52" Nov 29 05:13:43 crc kubenswrapper[4799]: I1129 05:13:43.301231 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/86cf64fe-3c75-4b05-8503-a4e3f3e0395c-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-x4w52\" (UID: \"86cf64fe-3c75-4b05-8503-a4e3f3e0395c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x4w52" Nov 29 05:13:43 crc kubenswrapper[4799]: I1129 05:13:43.301286 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/86cf64fe-3c75-4b05-8503-a4e3f3e0395c-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-x4w52\" (UID: \"86cf64fe-3c75-4b05-8503-a4e3f3e0395c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x4w52" Nov 29 05:13:43 crc kubenswrapper[4799]: I1129 05:13:43.403200 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/86cf64fe-3c75-4b05-8503-a4e3f3e0395c-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-x4w52\" (UID: \"86cf64fe-3c75-4b05-8503-a4e3f3e0395c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x4w52" Nov 29 05:13:43 crc kubenswrapper[4799]: I1129 05:13:43.403359 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/86cf64fe-3c75-4b05-8503-a4e3f3e0395c-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-x4w52\" (UID: \"86cf64fe-3c75-4b05-8503-a4e3f3e0395c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x4w52" Nov 29 05:13:43 crc kubenswrapper[4799]: I1129 05:13:43.403454 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/86cf64fe-3c75-4b05-8503-a4e3f3e0395c-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-x4w52\" (UID: \"86cf64fe-3c75-4b05-8503-a4e3f3e0395c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x4w52" Nov 29 05:13:43 crc kubenswrapper[4799]: I1129 05:13:43.403715 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ng8tz\" (UniqueName: \"kubernetes.io/projected/86cf64fe-3c75-4b05-8503-a4e3f3e0395c-kube-api-access-ng8tz\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-x4w52\" (UID: \"86cf64fe-3c75-4b05-8503-a4e3f3e0395c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x4w52" Nov 29 05:13:43 crc kubenswrapper[4799]: I1129 05:13:43.410938 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/86cf64fe-3c75-4b05-8503-a4e3f3e0395c-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-x4w52\" (UID: \"86cf64fe-3c75-4b05-8503-a4e3f3e0395c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x4w52" Nov 29 05:13:43 crc kubenswrapper[4799]: I1129 05:13:43.412253 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/86cf64fe-3c75-4b05-8503-a4e3f3e0395c-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-x4w52\" (UID: \"86cf64fe-3c75-4b05-8503-a4e3f3e0395c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x4w52" Nov 29 05:13:43 crc kubenswrapper[4799]: I1129 05:13:43.413393 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/86cf64fe-3c75-4b05-8503-a4e3f3e0395c-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-x4w52\" (UID: \"86cf64fe-3c75-4b05-8503-a4e3f3e0395c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x4w52" Nov 29 05:13:43 crc kubenswrapper[4799]: I1129 05:13:43.432672 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ng8tz\" (UniqueName: \"kubernetes.io/projected/86cf64fe-3c75-4b05-8503-a4e3f3e0395c-kube-api-access-ng8tz\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-x4w52\" (UID: \"86cf64fe-3c75-4b05-8503-a4e3f3e0395c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x4w52" Nov 29 05:13:43 crc kubenswrapper[4799]: I1129 05:13:43.477567 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x4w52" Nov 29 05:13:44 crc kubenswrapper[4799]: I1129 05:13:44.174241 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x4w52"] Nov 29 05:13:44 crc kubenswrapper[4799]: I1129 05:13:44.846174 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 29 05:13:45 crc kubenswrapper[4799]: I1129 05:13:45.006056 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x4w52" event={"ID":"86cf64fe-3c75-4b05-8503-a4e3f3e0395c","Type":"ContainerStarted","Data":"899c82e78c9efec57666fac0bbdecc1db4ea118c7a54add399dee0af7466961c"} Nov 29 05:13:46 crc kubenswrapper[4799]: I1129 05:13:46.030871 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x4w52" event={"ID":"86cf64fe-3c75-4b05-8503-a4e3f3e0395c","Type":"ContainerStarted","Data":"771d248c1bedbf58513295f96043db308918e610e30af14948da51131718293a"} Nov 29 05:13:46 crc kubenswrapper[4799]: I1129 05:13:46.063182 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x4w52" podStartSLOduration=2.409306223 podStartE2EDuration="3.063150496s" podCreationTimestamp="2025-11-29 05:13:43 +0000 UTC" firstStartedPulling="2025-11-29 05:13:44.186871133 +0000 UTC m=+2099.829801533" lastFinishedPulling="2025-11-29 05:13:44.840715396 +0000 UTC m=+2100.483645806" observedRunningTime="2025-11-29 05:13:46.055721816 +0000 UTC m=+2101.698652266" watchObservedRunningTime="2025-11-29 05:13:46.063150496 +0000 UTC m=+2101.706080926" Nov 29 05:14:12 crc kubenswrapper[4799]: I1129 05:14:12.699443 4799 generic.go:334] "Generic (PLEG): container finished" podID="86cf64fe-3c75-4b05-8503-a4e3f3e0395c" containerID="771d248c1bedbf58513295f96043db308918e610e30af14948da51131718293a" exitCode=0 Nov 29 05:14:12 crc kubenswrapper[4799]: I1129 05:14:12.699579 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x4w52" event={"ID":"86cf64fe-3c75-4b05-8503-a4e3f3e0395c","Type":"ContainerDied","Data":"771d248c1bedbf58513295f96043db308918e610e30af14948da51131718293a"} Nov 29 05:14:14 crc kubenswrapper[4799]: I1129 05:14:14.229765 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x4w52" Nov 29 05:14:14 crc kubenswrapper[4799]: I1129 05:14:14.354497 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/86cf64fe-3c75-4b05-8503-a4e3f3e0395c-ceph\") pod \"86cf64fe-3c75-4b05-8503-a4e3f3e0395c\" (UID: \"86cf64fe-3c75-4b05-8503-a4e3f3e0395c\") " Nov 29 05:14:14 crc kubenswrapper[4799]: I1129 05:14:14.354584 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/86cf64fe-3c75-4b05-8503-a4e3f3e0395c-ssh-key\") pod \"86cf64fe-3c75-4b05-8503-a4e3f3e0395c\" (UID: \"86cf64fe-3c75-4b05-8503-a4e3f3e0395c\") " Nov 29 05:14:14 crc kubenswrapper[4799]: I1129 05:14:14.354818 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ng8tz\" (UniqueName: \"kubernetes.io/projected/86cf64fe-3c75-4b05-8503-a4e3f3e0395c-kube-api-access-ng8tz\") pod \"86cf64fe-3c75-4b05-8503-a4e3f3e0395c\" (UID: \"86cf64fe-3c75-4b05-8503-a4e3f3e0395c\") " Nov 29 05:14:14 crc kubenswrapper[4799]: I1129 05:14:14.354909 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/86cf64fe-3c75-4b05-8503-a4e3f3e0395c-inventory\") pod \"86cf64fe-3c75-4b05-8503-a4e3f3e0395c\" (UID: \"86cf64fe-3c75-4b05-8503-a4e3f3e0395c\") " Nov 29 05:14:14 crc kubenswrapper[4799]: I1129 05:14:14.363525 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86cf64fe-3c75-4b05-8503-a4e3f3e0395c-kube-api-access-ng8tz" (OuterVolumeSpecName: "kube-api-access-ng8tz") pod "86cf64fe-3c75-4b05-8503-a4e3f3e0395c" (UID: "86cf64fe-3c75-4b05-8503-a4e3f3e0395c"). InnerVolumeSpecName "kube-api-access-ng8tz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:14:14 crc kubenswrapper[4799]: I1129 05:14:14.371251 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86cf64fe-3c75-4b05-8503-a4e3f3e0395c-ceph" (OuterVolumeSpecName: "ceph") pod "86cf64fe-3c75-4b05-8503-a4e3f3e0395c" (UID: "86cf64fe-3c75-4b05-8503-a4e3f3e0395c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:14:14 crc kubenswrapper[4799]: I1129 05:14:14.391996 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86cf64fe-3c75-4b05-8503-a4e3f3e0395c-inventory" (OuterVolumeSpecName: "inventory") pod "86cf64fe-3c75-4b05-8503-a4e3f3e0395c" (UID: "86cf64fe-3c75-4b05-8503-a4e3f3e0395c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:14:14 crc kubenswrapper[4799]: I1129 05:14:14.400071 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86cf64fe-3c75-4b05-8503-a4e3f3e0395c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "86cf64fe-3c75-4b05-8503-a4e3f3e0395c" (UID: "86cf64fe-3c75-4b05-8503-a4e3f3e0395c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:14:14 crc kubenswrapper[4799]: I1129 05:14:14.458133 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/86cf64fe-3c75-4b05-8503-a4e3f3e0395c-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 29 05:14:14 crc kubenswrapper[4799]: I1129 05:14:14.458180 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ng8tz\" (UniqueName: \"kubernetes.io/projected/86cf64fe-3c75-4b05-8503-a4e3f3e0395c-kube-api-access-ng8tz\") on node \"crc\" DevicePath \"\"" Nov 29 05:14:14 crc kubenswrapper[4799]: I1129 05:14:14.458195 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/86cf64fe-3c75-4b05-8503-a4e3f3e0395c-inventory\") on node \"crc\" DevicePath \"\"" Nov 29 05:14:14 crc kubenswrapper[4799]: I1129 05:14:14.458212 4799 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/86cf64fe-3c75-4b05-8503-a4e3f3e0395c-ceph\") on node \"crc\" DevicePath \"\"" Nov 29 05:14:14 crc kubenswrapper[4799]: I1129 05:14:14.725177 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x4w52" event={"ID":"86cf64fe-3c75-4b05-8503-a4e3f3e0395c","Type":"ContainerDied","Data":"899c82e78c9efec57666fac0bbdecc1db4ea118c7a54add399dee0af7466961c"} Nov 29 05:14:14 crc kubenswrapper[4799]: I1129 05:14:14.725259 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="899c82e78c9efec57666fac0bbdecc1db4ea118c7a54add399dee0af7466961c" Nov 29 05:14:14 crc kubenswrapper[4799]: I1129 05:14:14.725333 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x4w52" Nov 29 05:14:14 crc kubenswrapper[4799]: I1129 05:14:14.852318 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gnq8m"] Nov 29 05:14:14 crc kubenswrapper[4799]: E1129 05:14:14.852851 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86cf64fe-3c75-4b05-8503-a4e3f3e0395c" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 29 05:14:14 crc kubenswrapper[4799]: I1129 05:14:14.852872 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="86cf64fe-3c75-4b05-8503-a4e3f3e0395c" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 29 05:14:14 crc kubenswrapper[4799]: I1129 05:14:14.853047 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="86cf64fe-3c75-4b05-8503-a4e3f3e0395c" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 29 05:14:14 crc kubenswrapper[4799]: I1129 05:14:14.853819 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gnq8m" Nov 29 05:14:14 crc kubenswrapper[4799]: I1129 05:14:14.857625 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 29 05:14:14 crc kubenswrapper[4799]: I1129 05:14:14.857834 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 29 05:14:14 crc kubenswrapper[4799]: I1129 05:14:14.857889 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 29 05:14:14 crc kubenswrapper[4799]: I1129 05:14:14.857982 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ljwtk" Nov 29 05:14:14 crc kubenswrapper[4799]: I1129 05:14:14.858210 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 29 05:14:14 crc kubenswrapper[4799]: I1129 05:14:14.863948 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gnq8m"] Nov 29 05:14:14 crc kubenswrapper[4799]: I1129 05:14:14.971838 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/91c293ad-ab01-40e7-9fb6-b8ef99152e62-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-gnq8m\" (UID: \"91c293ad-ab01-40e7-9fb6-b8ef99152e62\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gnq8m" Nov 29 05:14:14 crc kubenswrapper[4799]: I1129 05:14:14.972335 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/91c293ad-ab01-40e7-9fb6-b8ef99152e62-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-gnq8m\" (UID: \"91c293ad-ab01-40e7-9fb6-b8ef99152e62\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gnq8m" Nov 29 05:14:14 crc kubenswrapper[4799]: I1129 05:14:14.972484 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/91c293ad-ab01-40e7-9fb6-b8ef99152e62-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-gnq8m\" (UID: \"91c293ad-ab01-40e7-9fb6-b8ef99152e62\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gnq8m" Nov 29 05:14:14 crc kubenswrapper[4799]: I1129 05:14:14.972682 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzvl2\" (UniqueName: \"kubernetes.io/projected/91c293ad-ab01-40e7-9fb6-b8ef99152e62-kube-api-access-wzvl2\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-gnq8m\" (UID: \"91c293ad-ab01-40e7-9fb6-b8ef99152e62\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gnq8m" Nov 29 05:14:15 crc kubenswrapper[4799]: I1129 05:14:15.074837 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/91c293ad-ab01-40e7-9fb6-b8ef99152e62-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-gnq8m\" (UID: \"91c293ad-ab01-40e7-9fb6-b8ef99152e62\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gnq8m" Nov 29 05:14:15 crc kubenswrapper[4799]: I1129 05:14:15.074941 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/91c293ad-ab01-40e7-9fb6-b8ef99152e62-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-gnq8m\" (UID: \"91c293ad-ab01-40e7-9fb6-b8ef99152e62\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gnq8m" Nov 29 05:14:15 crc kubenswrapper[4799]: I1129 05:14:15.075044 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzvl2\" (UniqueName: \"kubernetes.io/projected/91c293ad-ab01-40e7-9fb6-b8ef99152e62-kube-api-access-wzvl2\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-gnq8m\" (UID: \"91c293ad-ab01-40e7-9fb6-b8ef99152e62\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gnq8m" Nov 29 05:14:15 crc kubenswrapper[4799]: I1129 05:14:15.075181 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/91c293ad-ab01-40e7-9fb6-b8ef99152e62-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-gnq8m\" (UID: \"91c293ad-ab01-40e7-9fb6-b8ef99152e62\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gnq8m" Nov 29 05:14:15 crc kubenswrapper[4799]: I1129 05:14:15.081226 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/91c293ad-ab01-40e7-9fb6-b8ef99152e62-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-gnq8m\" (UID: \"91c293ad-ab01-40e7-9fb6-b8ef99152e62\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gnq8m" Nov 29 05:14:15 crc kubenswrapper[4799]: I1129 05:14:15.081230 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/91c293ad-ab01-40e7-9fb6-b8ef99152e62-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-gnq8m\" (UID: \"91c293ad-ab01-40e7-9fb6-b8ef99152e62\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gnq8m" Nov 29 05:14:15 crc kubenswrapper[4799]: I1129 05:14:15.081471 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/91c293ad-ab01-40e7-9fb6-b8ef99152e62-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-gnq8m\" (UID: \"91c293ad-ab01-40e7-9fb6-b8ef99152e62\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gnq8m" Nov 29 05:14:15 crc kubenswrapper[4799]: I1129 05:14:15.100535 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzvl2\" (UniqueName: \"kubernetes.io/projected/91c293ad-ab01-40e7-9fb6-b8ef99152e62-kube-api-access-wzvl2\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-gnq8m\" (UID: \"91c293ad-ab01-40e7-9fb6-b8ef99152e62\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gnq8m" Nov 29 05:14:15 crc kubenswrapper[4799]: I1129 05:14:15.186548 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gnq8m" Nov 29 05:14:15 crc kubenswrapper[4799]: I1129 05:14:15.876234 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gnq8m"] Nov 29 05:14:16 crc kubenswrapper[4799]: I1129 05:14:16.753482 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gnq8m" event={"ID":"91c293ad-ab01-40e7-9fb6-b8ef99152e62","Type":"ContainerStarted","Data":"1be54447901aed28bab4e85375a38c1111c6537d3fb3b6b5fc8040923ce15d55"} Nov 29 05:14:17 crc kubenswrapper[4799]: I1129 05:14:17.774087 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gnq8m" event={"ID":"91c293ad-ab01-40e7-9fb6-b8ef99152e62","Type":"ContainerStarted","Data":"25916f3d6ac2216644a7f24449d36aa955c883844cca6201f51a773874a3cff6"} Nov 29 05:14:17 crc kubenswrapper[4799]: I1129 05:14:17.805100 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gnq8m" podStartSLOduration=3.067046424 podStartE2EDuration="3.805063164s" podCreationTimestamp="2025-11-29 05:14:14 +0000 UTC" firstStartedPulling="2025-11-29 05:14:15.895968717 +0000 UTC m=+2131.538899127" lastFinishedPulling="2025-11-29 05:14:16.633985437 +0000 UTC m=+2132.276915867" observedRunningTime="2025-11-29 05:14:17.797228855 +0000 UTC m=+2133.440159295" watchObservedRunningTime="2025-11-29 05:14:17.805063164 +0000 UTC m=+2133.447993594" Nov 29 05:14:22 crc kubenswrapper[4799]: I1129 05:14:22.833410 4799 generic.go:334] "Generic (PLEG): container finished" podID="91c293ad-ab01-40e7-9fb6-b8ef99152e62" containerID="25916f3d6ac2216644a7f24449d36aa955c883844cca6201f51a773874a3cff6" exitCode=0 Nov 29 05:14:22 crc kubenswrapper[4799]: I1129 05:14:22.833544 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gnq8m" event={"ID":"91c293ad-ab01-40e7-9fb6-b8ef99152e62","Type":"ContainerDied","Data":"25916f3d6ac2216644a7f24449d36aa955c883844cca6201f51a773874a3cff6"} Nov 29 05:14:24 crc kubenswrapper[4799]: I1129 05:14:24.353047 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gnq8m" Nov 29 05:14:24 crc kubenswrapper[4799]: I1129 05:14:24.417524 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/91c293ad-ab01-40e7-9fb6-b8ef99152e62-ceph\") pod \"91c293ad-ab01-40e7-9fb6-b8ef99152e62\" (UID: \"91c293ad-ab01-40e7-9fb6-b8ef99152e62\") " Nov 29 05:14:24 crc kubenswrapper[4799]: I1129 05:14:24.417767 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wzvl2\" (UniqueName: \"kubernetes.io/projected/91c293ad-ab01-40e7-9fb6-b8ef99152e62-kube-api-access-wzvl2\") pod \"91c293ad-ab01-40e7-9fb6-b8ef99152e62\" (UID: \"91c293ad-ab01-40e7-9fb6-b8ef99152e62\") " Nov 29 05:14:24 crc kubenswrapper[4799]: I1129 05:14:24.417829 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/91c293ad-ab01-40e7-9fb6-b8ef99152e62-ssh-key\") pod \"91c293ad-ab01-40e7-9fb6-b8ef99152e62\" (UID: \"91c293ad-ab01-40e7-9fb6-b8ef99152e62\") " Nov 29 05:14:24 crc kubenswrapper[4799]: I1129 05:14:24.418086 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/91c293ad-ab01-40e7-9fb6-b8ef99152e62-inventory\") pod \"91c293ad-ab01-40e7-9fb6-b8ef99152e62\" (UID: \"91c293ad-ab01-40e7-9fb6-b8ef99152e62\") " Nov 29 05:14:24 crc kubenswrapper[4799]: I1129 05:14:24.426282 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91c293ad-ab01-40e7-9fb6-b8ef99152e62-ceph" (OuterVolumeSpecName: "ceph") pod "91c293ad-ab01-40e7-9fb6-b8ef99152e62" (UID: "91c293ad-ab01-40e7-9fb6-b8ef99152e62"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:14:24 crc kubenswrapper[4799]: I1129 05:14:24.435344 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91c293ad-ab01-40e7-9fb6-b8ef99152e62-kube-api-access-wzvl2" (OuterVolumeSpecName: "kube-api-access-wzvl2") pod "91c293ad-ab01-40e7-9fb6-b8ef99152e62" (UID: "91c293ad-ab01-40e7-9fb6-b8ef99152e62"). InnerVolumeSpecName "kube-api-access-wzvl2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:14:24 crc kubenswrapper[4799]: I1129 05:14:24.457291 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91c293ad-ab01-40e7-9fb6-b8ef99152e62-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "91c293ad-ab01-40e7-9fb6-b8ef99152e62" (UID: "91c293ad-ab01-40e7-9fb6-b8ef99152e62"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:14:24 crc kubenswrapper[4799]: I1129 05:14:24.473486 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91c293ad-ab01-40e7-9fb6-b8ef99152e62-inventory" (OuterVolumeSpecName: "inventory") pod "91c293ad-ab01-40e7-9fb6-b8ef99152e62" (UID: "91c293ad-ab01-40e7-9fb6-b8ef99152e62"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:14:24 crc kubenswrapper[4799]: I1129 05:14:24.521118 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wzvl2\" (UniqueName: \"kubernetes.io/projected/91c293ad-ab01-40e7-9fb6-b8ef99152e62-kube-api-access-wzvl2\") on node \"crc\" DevicePath \"\"" Nov 29 05:14:24 crc kubenswrapper[4799]: I1129 05:14:24.521179 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/91c293ad-ab01-40e7-9fb6-b8ef99152e62-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 29 05:14:24 crc kubenswrapper[4799]: I1129 05:14:24.521196 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/91c293ad-ab01-40e7-9fb6-b8ef99152e62-inventory\") on node \"crc\" DevicePath \"\"" Nov 29 05:14:24 crc kubenswrapper[4799]: I1129 05:14:24.521214 4799 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/91c293ad-ab01-40e7-9fb6-b8ef99152e62-ceph\") on node \"crc\" DevicePath \"\"" Nov 29 05:14:24 crc kubenswrapper[4799]: I1129 05:14:24.863491 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gnq8m" event={"ID":"91c293ad-ab01-40e7-9fb6-b8ef99152e62","Type":"ContainerDied","Data":"1be54447901aed28bab4e85375a38c1111c6537d3fb3b6b5fc8040923ce15d55"} Nov 29 05:14:24 crc kubenswrapper[4799]: I1129 05:14:24.864183 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1be54447901aed28bab4e85375a38c1111c6537d3fb3b6b5fc8040923ce15d55" Nov 29 05:14:24 crc kubenswrapper[4799]: I1129 05:14:24.863614 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gnq8m" Nov 29 05:14:24 crc kubenswrapper[4799]: I1129 05:14:24.946983 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-hdkhl"] Nov 29 05:14:24 crc kubenswrapper[4799]: E1129 05:14:24.947636 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91c293ad-ab01-40e7-9fb6-b8ef99152e62" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 29 05:14:24 crc kubenswrapper[4799]: I1129 05:14:24.947664 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="91c293ad-ab01-40e7-9fb6-b8ef99152e62" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 29 05:14:24 crc kubenswrapper[4799]: I1129 05:14:24.947947 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="91c293ad-ab01-40e7-9fb6-b8ef99152e62" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 29 05:14:24 crc kubenswrapper[4799]: I1129 05:14:24.948884 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hdkhl" Nov 29 05:14:24 crc kubenswrapper[4799]: I1129 05:14:24.951329 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 29 05:14:24 crc kubenswrapper[4799]: I1129 05:14:24.951477 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 29 05:14:24 crc kubenswrapper[4799]: I1129 05:14:24.951627 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ljwtk" Nov 29 05:14:24 crc kubenswrapper[4799]: I1129 05:14:24.952025 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 29 05:14:24 crc kubenswrapper[4799]: I1129 05:14:24.956050 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-hdkhl"] Nov 29 05:14:24 crc kubenswrapper[4799]: I1129 05:14:24.961054 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 29 05:14:25 crc kubenswrapper[4799]: I1129 05:14:25.034415 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nmjq\" (UniqueName: \"kubernetes.io/projected/1784d248-9169-4b88-b8d1-16412106a8dc-kube-api-access-2nmjq\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hdkhl\" (UID: \"1784d248-9169-4b88-b8d1-16412106a8dc\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hdkhl" Nov 29 05:14:25 crc kubenswrapper[4799]: I1129 05:14:25.034495 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1784d248-9169-4b88-b8d1-16412106a8dc-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hdkhl\" (UID: \"1784d248-9169-4b88-b8d1-16412106a8dc\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hdkhl" Nov 29 05:14:25 crc kubenswrapper[4799]: I1129 05:14:25.034589 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1784d248-9169-4b88-b8d1-16412106a8dc-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hdkhl\" (UID: \"1784d248-9169-4b88-b8d1-16412106a8dc\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hdkhl" Nov 29 05:14:25 crc kubenswrapper[4799]: I1129 05:14:25.034639 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1784d248-9169-4b88-b8d1-16412106a8dc-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hdkhl\" (UID: \"1784d248-9169-4b88-b8d1-16412106a8dc\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hdkhl" Nov 29 05:14:25 crc kubenswrapper[4799]: I1129 05:14:25.136911 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nmjq\" (UniqueName: \"kubernetes.io/projected/1784d248-9169-4b88-b8d1-16412106a8dc-kube-api-access-2nmjq\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hdkhl\" (UID: \"1784d248-9169-4b88-b8d1-16412106a8dc\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hdkhl" Nov 29 05:14:25 crc kubenswrapper[4799]: I1129 05:14:25.136965 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1784d248-9169-4b88-b8d1-16412106a8dc-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hdkhl\" (UID: \"1784d248-9169-4b88-b8d1-16412106a8dc\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hdkhl" Nov 29 05:14:25 crc kubenswrapper[4799]: I1129 05:14:25.137025 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1784d248-9169-4b88-b8d1-16412106a8dc-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hdkhl\" (UID: \"1784d248-9169-4b88-b8d1-16412106a8dc\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hdkhl" Nov 29 05:14:25 crc kubenswrapper[4799]: I1129 05:14:25.137072 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1784d248-9169-4b88-b8d1-16412106a8dc-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hdkhl\" (UID: \"1784d248-9169-4b88-b8d1-16412106a8dc\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hdkhl" Nov 29 05:14:25 crc kubenswrapper[4799]: I1129 05:14:25.143023 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1784d248-9169-4b88-b8d1-16412106a8dc-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hdkhl\" (UID: \"1784d248-9169-4b88-b8d1-16412106a8dc\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hdkhl" Nov 29 05:14:25 crc kubenswrapper[4799]: I1129 05:14:25.143023 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1784d248-9169-4b88-b8d1-16412106a8dc-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hdkhl\" (UID: \"1784d248-9169-4b88-b8d1-16412106a8dc\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hdkhl" Nov 29 05:14:25 crc kubenswrapper[4799]: I1129 05:14:25.143139 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1784d248-9169-4b88-b8d1-16412106a8dc-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hdkhl\" (UID: \"1784d248-9169-4b88-b8d1-16412106a8dc\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hdkhl" Nov 29 05:14:25 crc kubenswrapper[4799]: I1129 05:14:25.160245 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nmjq\" (UniqueName: \"kubernetes.io/projected/1784d248-9169-4b88-b8d1-16412106a8dc-kube-api-access-2nmjq\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hdkhl\" (UID: \"1784d248-9169-4b88-b8d1-16412106a8dc\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hdkhl" Nov 29 05:14:25 crc kubenswrapper[4799]: I1129 05:14:25.273040 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hdkhl" Nov 29 05:14:25 crc kubenswrapper[4799]: I1129 05:14:25.739756 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-hdkhl"] Nov 29 05:14:25 crc kubenswrapper[4799]: I1129 05:14:25.873272 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hdkhl" event={"ID":"1784d248-9169-4b88-b8d1-16412106a8dc","Type":"ContainerStarted","Data":"c8f57d315c0cdff4cd8c349721765bcbd76ad5741881e3d5bf5c77517ecbfa94"} Nov 29 05:14:26 crc kubenswrapper[4799]: I1129 05:14:26.896687 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hdkhl" event={"ID":"1784d248-9169-4b88-b8d1-16412106a8dc","Type":"ContainerStarted","Data":"9117378b7ca1228317225c3a69913f0cdcb9c6095b79838cfe888a8c39ba5c0f"} Nov 29 05:14:26 crc kubenswrapper[4799]: I1129 05:14:26.937510 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hdkhl" podStartSLOduration=2.4907799170000002 podStartE2EDuration="2.937476619s" podCreationTimestamp="2025-11-29 05:14:24 +0000 UTC" firstStartedPulling="2025-11-29 05:14:25.752763649 +0000 UTC m=+2141.395694049" lastFinishedPulling="2025-11-29 05:14:26.199460341 +0000 UTC m=+2141.842390751" observedRunningTime="2025-11-29 05:14:26.917000967 +0000 UTC m=+2142.559931387" watchObservedRunningTime="2025-11-29 05:14:26.937476619 +0000 UTC m=+2142.580407039" Nov 29 05:15:00 crc kubenswrapper[4799]: I1129 05:15:00.204565 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406555-g8t85"] Nov 29 05:15:00 crc kubenswrapper[4799]: I1129 05:15:00.207309 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406555-g8t85" Nov 29 05:15:00 crc kubenswrapper[4799]: I1129 05:15:00.210895 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 29 05:15:00 crc kubenswrapper[4799]: I1129 05:15:00.212486 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 29 05:15:00 crc kubenswrapper[4799]: I1129 05:15:00.216417 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406555-g8t85"] Nov 29 05:15:00 crc kubenswrapper[4799]: I1129 05:15:00.306707 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a2dad815-9bd5-4c9f-957a-16ed0df1ff5e-secret-volume\") pod \"collect-profiles-29406555-g8t85\" (UID: \"a2dad815-9bd5-4c9f-957a-16ed0df1ff5e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406555-g8t85" Nov 29 05:15:00 crc kubenswrapper[4799]: I1129 05:15:00.307313 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a2dad815-9bd5-4c9f-957a-16ed0df1ff5e-config-volume\") pod \"collect-profiles-29406555-g8t85\" (UID: \"a2dad815-9bd5-4c9f-957a-16ed0df1ff5e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406555-g8t85" Nov 29 05:15:00 crc kubenswrapper[4799]: I1129 05:15:00.307438 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tw4fj\" (UniqueName: \"kubernetes.io/projected/a2dad815-9bd5-4c9f-957a-16ed0df1ff5e-kube-api-access-tw4fj\") pod \"collect-profiles-29406555-g8t85\" (UID: \"a2dad815-9bd5-4c9f-957a-16ed0df1ff5e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406555-g8t85" Nov 29 05:15:00 crc kubenswrapper[4799]: I1129 05:15:00.410730 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a2dad815-9bd5-4c9f-957a-16ed0df1ff5e-config-volume\") pod \"collect-profiles-29406555-g8t85\" (UID: \"a2dad815-9bd5-4c9f-957a-16ed0df1ff5e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406555-g8t85" Nov 29 05:15:00 crc kubenswrapper[4799]: I1129 05:15:00.410829 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tw4fj\" (UniqueName: \"kubernetes.io/projected/a2dad815-9bd5-4c9f-957a-16ed0df1ff5e-kube-api-access-tw4fj\") pod \"collect-profiles-29406555-g8t85\" (UID: \"a2dad815-9bd5-4c9f-957a-16ed0df1ff5e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406555-g8t85" Nov 29 05:15:00 crc kubenswrapper[4799]: I1129 05:15:00.410932 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a2dad815-9bd5-4c9f-957a-16ed0df1ff5e-secret-volume\") pod \"collect-profiles-29406555-g8t85\" (UID: \"a2dad815-9bd5-4c9f-957a-16ed0df1ff5e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406555-g8t85" Nov 29 05:15:00 crc kubenswrapper[4799]: I1129 05:15:00.411818 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a2dad815-9bd5-4c9f-957a-16ed0df1ff5e-config-volume\") pod \"collect-profiles-29406555-g8t85\" (UID: \"a2dad815-9bd5-4c9f-957a-16ed0df1ff5e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406555-g8t85" Nov 29 05:15:00 crc kubenswrapper[4799]: I1129 05:15:00.419729 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a2dad815-9bd5-4c9f-957a-16ed0df1ff5e-secret-volume\") pod \"collect-profiles-29406555-g8t85\" (UID: \"a2dad815-9bd5-4c9f-957a-16ed0df1ff5e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406555-g8t85" Nov 29 05:15:00 crc kubenswrapper[4799]: I1129 05:15:00.441625 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tw4fj\" (UniqueName: \"kubernetes.io/projected/a2dad815-9bd5-4c9f-957a-16ed0df1ff5e-kube-api-access-tw4fj\") pod \"collect-profiles-29406555-g8t85\" (UID: \"a2dad815-9bd5-4c9f-957a-16ed0df1ff5e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406555-g8t85" Nov 29 05:15:00 crc kubenswrapper[4799]: I1129 05:15:00.529628 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406555-g8t85" Nov 29 05:15:00 crc kubenswrapper[4799]: I1129 05:15:00.840086 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406555-g8t85"] Nov 29 05:15:01 crc kubenswrapper[4799]: I1129 05:15:01.279652 4799 generic.go:334] "Generic (PLEG): container finished" podID="a2dad815-9bd5-4c9f-957a-16ed0df1ff5e" containerID="fe872e5362e75db2c42487603d95f07dc9590830840138f3bf0c3f0d4085ae2c" exitCode=0 Nov 29 05:15:01 crc kubenswrapper[4799]: I1129 05:15:01.279723 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29406555-g8t85" event={"ID":"a2dad815-9bd5-4c9f-957a-16ed0df1ff5e","Type":"ContainerDied","Data":"fe872e5362e75db2c42487603d95f07dc9590830840138f3bf0c3f0d4085ae2c"} Nov 29 05:15:01 crc kubenswrapper[4799]: I1129 05:15:01.279816 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29406555-g8t85" event={"ID":"a2dad815-9bd5-4c9f-957a-16ed0df1ff5e","Type":"ContainerStarted","Data":"44dc53ce152fa4e9bc0d970cb5d59fb53308a0d418e95781221a9185dc05d5af"} Nov 29 05:15:02 crc kubenswrapper[4799]: I1129 05:15:02.711535 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406555-g8t85" Nov 29 05:15:02 crc kubenswrapper[4799]: I1129 05:15:02.765742 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a2dad815-9bd5-4c9f-957a-16ed0df1ff5e-secret-volume\") pod \"a2dad815-9bd5-4c9f-957a-16ed0df1ff5e\" (UID: \"a2dad815-9bd5-4c9f-957a-16ed0df1ff5e\") " Nov 29 05:15:02 crc kubenswrapper[4799]: I1129 05:15:02.765879 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tw4fj\" (UniqueName: \"kubernetes.io/projected/a2dad815-9bd5-4c9f-957a-16ed0df1ff5e-kube-api-access-tw4fj\") pod \"a2dad815-9bd5-4c9f-957a-16ed0df1ff5e\" (UID: \"a2dad815-9bd5-4c9f-957a-16ed0df1ff5e\") " Nov 29 05:15:02 crc kubenswrapper[4799]: I1129 05:15:02.765921 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a2dad815-9bd5-4c9f-957a-16ed0df1ff5e-config-volume\") pod \"a2dad815-9bd5-4c9f-957a-16ed0df1ff5e\" (UID: \"a2dad815-9bd5-4c9f-957a-16ed0df1ff5e\") " Nov 29 05:15:02 crc kubenswrapper[4799]: I1129 05:15:02.767391 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2dad815-9bd5-4c9f-957a-16ed0df1ff5e-config-volume" (OuterVolumeSpecName: "config-volume") pod "a2dad815-9bd5-4c9f-957a-16ed0df1ff5e" (UID: "a2dad815-9bd5-4c9f-957a-16ed0df1ff5e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 05:15:02 crc kubenswrapper[4799]: I1129 05:15:02.773961 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2dad815-9bd5-4c9f-957a-16ed0df1ff5e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a2dad815-9bd5-4c9f-957a-16ed0df1ff5e" (UID: "a2dad815-9bd5-4c9f-957a-16ed0df1ff5e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:15:02 crc kubenswrapper[4799]: I1129 05:15:02.775202 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2dad815-9bd5-4c9f-957a-16ed0df1ff5e-kube-api-access-tw4fj" (OuterVolumeSpecName: "kube-api-access-tw4fj") pod "a2dad815-9bd5-4c9f-957a-16ed0df1ff5e" (UID: "a2dad815-9bd5-4c9f-957a-16ed0df1ff5e"). InnerVolumeSpecName "kube-api-access-tw4fj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:15:02 crc kubenswrapper[4799]: I1129 05:15:02.869344 4799 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a2dad815-9bd5-4c9f-957a-16ed0df1ff5e-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 29 05:15:02 crc kubenswrapper[4799]: I1129 05:15:02.869394 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tw4fj\" (UniqueName: \"kubernetes.io/projected/a2dad815-9bd5-4c9f-957a-16ed0df1ff5e-kube-api-access-tw4fj\") on node \"crc\" DevicePath \"\"" Nov 29 05:15:02 crc kubenswrapper[4799]: I1129 05:15:02.869405 4799 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a2dad815-9bd5-4c9f-957a-16ed0df1ff5e-config-volume\") on node \"crc\" DevicePath \"\"" Nov 29 05:15:03 crc kubenswrapper[4799]: I1129 05:15:03.307338 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29406555-g8t85" event={"ID":"a2dad815-9bd5-4c9f-957a-16ed0df1ff5e","Type":"ContainerDied","Data":"44dc53ce152fa4e9bc0d970cb5d59fb53308a0d418e95781221a9185dc05d5af"} Nov 29 05:15:03 crc kubenswrapper[4799]: I1129 05:15:03.307395 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406555-g8t85" Nov 29 05:15:03 crc kubenswrapper[4799]: I1129 05:15:03.307417 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="44dc53ce152fa4e9bc0d970cb5d59fb53308a0d418e95781221a9185dc05d5af" Nov 29 05:15:03 crc kubenswrapper[4799]: I1129 05:15:03.806385 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406510-f2lz7"] Nov 29 05:15:03 crc kubenswrapper[4799]: I1129 05:15:03.819958 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406510-f2lz7"] Nov 29 05:15:04 crc kubenswrapper[4799]: I1129 05:15:04.674984 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96a1e39b-0018-41b8-8d7c-9fa7697858e9" path="/var/lib/kubelet/pods/96a1e39b-0018-41b8-8d7c-9fa7697858e9/volumes" Nov 29 05:15:07 crc kubenswrapper[4799]: I1129 05:15:07.497855 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 05:15:07 crc kubenswrapper[4799]: I1129 05:15:07.498350 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 05:15:07 crc kubenswrapper[4799]: E1129 05:15:07.700776 4799 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1784d248_9169_4b88_b8d1_16412106a8dc.slice/crio-conmon-9117378b7ca1228317225c3a69913f0cdcb9c6095b79838cfe888a8c39ba5c0f.scope\": RecentStats: unable to find data in memory cache]" Nov 29 05:15:08 crc kubenswrapper[4799]: I1129 05:15:08.370989 4799 generic.go:334] "Generic (PLEG): container finished" podID="1784d248-9169-4b88-b8d1-16412106a8dc" containerID="9117378b7ca1228317225c3a69913f0cdcb9c6095b79838cfe888a8c39ba5c0f" exitCode=0 Nov 29 05:15:08 crc kubenswrapper[4799]: I1129 05:15:08.371100 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hdkhl" event={"ID":"1784d248-9169-4b88-b8d1-16412106a8dc","Type":"ContainerDied","Data":"9117378b7ca1228317225c3a69913f0cdcb9c6095b79838cfe888a8c39ba5c0f"} Nov 29 05:15:09 crc kubenswrapper[4799]: I1129 05:15:09.906959 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hdkhl" Nov 29 05:15:09 crc kubenswrapper[4799]: I1129 05:15:09.947057 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1784d248-9169-4b88-b8d1-16412106a8dc-ssh-key\") pod \"1784d248-9169-4b88-b8d1-16412106a8dc\" (UID: \"1784d248-9169-4b88-b8d1-16412106a8dc\") " Nov 29 05:15:09 crc kubenswrapper[4799]: I1129 05:15:09.947130 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1784d248-9169-4b88-b8d1-16412106a8dc-inventory\") pod \"1784d248-9169-4b88-b8d1-16412106a8dc\" (UID: \"1784d248-9169-4b88-b8d1-16412106a8dc\") " Nov 29 05:15:09 crc kubenswrapper[4799]: I1129 05:15:09.947290 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1784d248-9169-4b88-b8d1-16412106a8dc-ceph\") pod \"1784d248-9169-4b88-b8d1-16412106a8dc\" (UID: \"1784d248-9169-4b88-b8d1-16412106a8dc\") " Nov 29 05:15:09 crc kubenswrapper[4799]: I1129 05:15:09.947442 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2nmjq\" (UniqueName: \"kubernetes.io/projected/1784d248-9169-4b88-b8d1-16412106a8dc-kube-api-access-2nmjq\") pod \"1784d248-9169-4b88-b8d1-16412106a8dc\" (UID: \"1784d248-9169-4b88-b8d1-16412106a8dc\") " Nov 29 05:15:09 crc kubenswrapper[4799]: I1129 05:15:09.960711 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1784d248-9169-4b88-b8d1-16412106a8dc-ceph" (OuterVolumeSpecName: "ceph") pod "1784d248-9169-4b88-b8d1-16412106a8dc" (UID: "1784d248-9169-4b88-b8d1-16412106a8dc"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:15:09 crc kubenswrapper[4799]: I1129 05:15:09.962510 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1784d248-9169-4b88-b8d1-16412106a8dc-kube-api-access-2nmjq" (OuterVolumeSpecName: "kube-api-access-2nmjq") pod "1784d248-9169-4b88-b8d1-16412106a8dc" (UID: "1784d248-9169-4b88-b8d1-16412106a8dc"). InnerVolumeSpecName "kube-api-access-2nmjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:15:09 crc kubenswrapper[4799]: I1129 05:15:09.996384 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1784d248-9169-4b88-b8d1-16412106a8dc-inventory" (OuterVolumeSpecName: "inventory") pod "1784d248-9169-4b88-b8d1-16412106a8dc" (UID: "1784d248-9169-4b88-b8d1-16412106a8dc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:15:10 crc kubenswrapper[4799]: I1129 05:15:10.001345 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1784d248-9169-4b88-b8d1-16412106a8dc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1784d248-9169-4b88-b8d1-16412106a8dc" (UID: "1784d248-9169-4b88-b8d1-16412106a8dc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:15:10 crc kubenswrapper[4799]: I1129 05:15:10.051904 4799 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1784d248-9169-4b88-b8d1-16412106a8dc-ceph\") on node \"crc\" DevicePath \"\"" Nov 29 05:15:10 crc kubenswrapper[4799]: I1129 05:15:10.051958 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2nmjq\" (UniqueName: \"kubernetes.io/projected/1784d248-9169-4b88-b8d1-16412106a8dc-kube-api-access-2nmjq\") on node \"crc\" DevicePath \"\"" Nov 29 05:15:10 crc kubenswrapper[4799]: I1129 05:15:10.051980 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1784d248-9169-4b88-b8d1-16412106a8dc-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 29 05:15:10 crc kubenswrapper[4799]: I1129 05:15:10.052003 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1784d248-9169-4b88-b8d1-16412106a8dc-inventory\") on node \"crc\" DevicePath \"\"" Nov 29 05:15:10 crc kubenswrapper[4799]: I1129 05:15:10.427356 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hdkhl" event={"ID":"1784d248-9169-4b88-b8d1-16412106a8dc","Type":"ContainerDied","Data":"c8f57d315c0cdff4cd8c349721765bcbd76ad5741881e3d5bf5c77517ecbfa94"} Nov 29 05:15:10 crc kubenswrapper[4799]: I1129 05:15:10.428070 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c8f57d315c0cdff4cd8c349721765bcbd76ad5741881e3d5bf5c77517ecbfa94" Nov 29 05:15:10 crc kubenswrapper[4799]: I1129 05:15:10.427623 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hdkhl" Nov 29 05:15:10 crc kubenswrapper[4799]: I1129 05:15:10.529955 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6rm8k"] Nov 29 05:15:10 crc kubenswrapper[4799]: E1129 05:15:10.530533 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1784d248-9169-4b88-b8d1-16412106a8dc" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 29 05:15:10 crc kubenswrapper[4799]: I1129 05:15:10.530568 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="1784d248-9169-4b88-b8d1-16412106a8dc" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 29 05:15:10 crc kubenswrapper[4799]: E1129 05:15:10.530611 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2dad815-9bd5-4c9f-957a-16ed0df1ff5e" containerName="collect-profiles" Nov 29 05:15:10 crc kubenswrapper[4799]: I1129 05:15:10.530625 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2dad815-9bd5-4c9f-957a-16ed0df1ff5e" containerName="collect-profiles" Nov 29 05:15:10 crc kubenswrapper[4799]: I1129 05:15:10.530972 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2dad815-9bd5-4c9f-957a-16ed0df1ff5e" containerName="collect-profiles" Nov 29 05:15:10 crc kubenswrapper[4799]: I1129 05:15:10.531021 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="1784d248-9169-4b88-b8d1-16412106a8dc" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 29 05:15:10 crc kubenswrapper[4799]: I1129 05:15:10.532010 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6rm8k" Nov 29 05:15:10 crc kubenswrapper[4799]: I1129 05:15:10.535479 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ljwtk" Nov 29 05:15:10 crc kubenswrapper[4799]: I1129 05:15:10.535670 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 29 05:15:10 crc kubenswrapper[4799]: I1129 05:15:10.536360 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 29 05:15:10 crc kubenswrapper[4799]: I1129 05:15:10.537141 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 29 05:15:10 crc kubenswrapper[4799]: I1129 05:15:10.545600 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 29 05:15:10 crc kubenswrapper[4799]: I1129 05:15:10.551287 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6rm8k"] Nov 29 05:15:10 crc kubenswrapper[4799]: I1129 05:15:10.561349 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/92efc83b-d2e4-44f7-8ede-db45686a6c0f-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6rm8k\" (UID: \"92efc83b-d2e4-44f7-8ede-db45686a6c0f\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6rm8k" Nov 29 05:15:10 crc kubenswrapper[4799]: I1129 05:15:10.561540 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kntp\" (UniqueName: \"kubernetes.io/projected/92efc83b-d2e4-44f7-8ede-db45686a6c0f-kube-api-access-7kntp\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6rm8k\" (UID: \"92efc83b-d2e4-44f7-8ede-db45686a6c0f\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6rm8k" Nov 29 05:15:10 crc kubenswrapper[4799]: I1129 05:15:10.561630 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/92efc83b-d2e4-44f7-8ede-db45686a6c0f-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6rm8k\" (UID: \"92efc83b-d2e4-44f7-8ede-db45686a6c0f\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6rm8k" Nov 29 05:15:10 crc kubenswrapper[4799]: I1129 05:15:10.561695 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/92efc83b-d2e4-44f7-8ede-db45686a6c0f-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6rm8k\" (UID: \"92efc83b-d2e4-44f7-8ede-db45686a6c0f\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6rm8k" Nov 29 05:15:10 crc kubenswrapper[4799]: I1129 05:15:10.664458 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/92efc83b-d2e4-44f7-8ede-db45686a6c0f-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6rm8k\" (UID: \"92efc83b-d2e4-44f7-8ede-db45686a6c0f\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6rm8k" Nov 29 05:15:10 crc kubenswrapper[4799]: I1129 05:15:10.664617 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kntp\" (UniqueName: \"kubernetes.io/projected/92efc83b-d2e4-44f7-8ede-db45686a6c0f-kube-api-access-7kntp\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6rm8k\" (UID: \"92efc83b-d2e4-44f7-8ede-db45686a6c0f\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6rm8k" Nov 29 05:15:10 crc kubenswrapper[4799]: I1129 05:15:10.664679 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/92efc83b-d2e4-44f7-8ede-db45686a6c0f-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6rm8k\" (UID: \"92efc83b-d2e4-44f7-8ede-db45686a6c0f\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6rm8k" Nov 29 05:15:10 crc kubenswrapper[4799]: I1129 05:15:10.664722 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/92efc83b-d2e4-44f7-8ede-db45686a6c0f-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6rm8k\" (UID: \"92efc83b-d2e4-44f7-8ede-db45686a6c0f\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6rm8k" Nov 29 05:15:10 crc kubenswrapper[4799]: I1129 05:15:10.669213 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/92efc83b-d2e4-44f7-8ede-db45686a6c0f-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6rm8k\" (UID: \"92efc83b-d2e4-44f7-8ede-db45686a6c0f\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6rm8k" Nov 29 05:15:10 crc kubenswrapper[4799]: I1129 05:15:10.672849 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/92efc83b-d2e4-44f7-8ede-db45686a6c0f-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6rm8k\" (UID: \"92efc83b-d2e4-44f7-8ede-db45686a6c0f\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6rm8k" Nov 29 05:15:10 crc kubenswrapper[4799]: I1129 05:15:10.685550 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/92efc83b-d2e4-44f7-8ede-db45686a6c0f-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6rm8k\" (UID: \"92efc83b-d2e4-44f7-8ede-db45686a6c0f\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6rm8k" Nov 29 05:15:10 crc kubenswrapper[4799]: I1129 05:15:10.694003 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kntp\" (UniqueName: \"kubernetes.io/projected/92efc83b-d2e4-44f7-8ede-db45686a6c0f-kube-api-access-7kntp\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6rm8k\" (UID: \"92efc83b-d2e4-44f7-8ede-db45686a6c0f\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6rm8k" Nov 29 05:15:10 crc kubenswrapper[4799]: I1129 05:15:10.859451 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6rm8k" Nov 29 05:15:11 crc kubenswrapper[4799]: I1129 05:15:11.299313 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6rm8k"] Nov 29 05:15:11 crc kubenswrapper[4799]: I1129 05:15:11.441567 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6rm8k" event={"ID":"92efc83b-d2e4-44f7-8ede-db45686a6c0f","Type":"ContainerStarted","Data":"2c1ca7ce78c1ba2d982ef5934b9d7a7be6d23a7b11a68cb72e748fad442cc516"} Nov 29 05:15:12 crc kubenswrapper[4799]: I1129 05:15:12.456506 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6rm8k" event={"ID":"92efc83b-d2e4-44f7-8ede-db45686a6c0f","Type":"ContainerStarted","Data":"059d74fc5ab3f32e2c9c104b552f40c0b550335427e8ac36026515c8854b5bfd"} Nov 29 05:15:12 crc kubenswrapper[4799]: I1129 05:15:12.488960 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6rm8k" podStartSLOduration=1.838052755 podStartE2EDuration="2.488928849s" podCreationTimestamp="2025-11-29 05:15:10 +0000 UTC" firstStartedPulling="2025-11-29 05:15:11.30788936 +0000 UTC m=+2186.950819760" lastFinishedPulling="2025-11-29 05:15:11.958765434 +0000 UTC m=+2187.601695854" observedRunningTime="2025-11-29 05:15:12.481538669 +0000 UTC m=+2188.124469109" watchObservedRunningTime="2025-11-29 05:15:12.488928849 +0000 UTC m=+2188.131859289" Nov 29 05:15:13 crc kubenswrapper[4799]: I1129 05:15:13.817216 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-scs7t"] Nov 29 05:15:13 crc kubenswrapper[4799]: I1129 05:15:13.822253 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-scs7t" Nov 29 05:15:13 crc kubenswrapper[4799]: I1129 05:15:13.843995 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-scs7t"] Nov 29 05:15:13 crc kubenswrapper[4799]: I1129 05:15:13.952369 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4990ec88-046b-4039-8ae8-062cf8c97b14-catalog-content\") pod \"redhat-marketplace-scs7t\" (UID: \"4990ec88-046b-4039-8ae8-062cf8c97b14\") " pod="openshift-marketplace/redhat-marketplace-scs7t" Nov 29 05:15:13 crc kubenswrapper[4799]: I1129 05:15:13.953767 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4990ec88-046b-4039-8ae8-062cf8c97b14-utilities\") pod \"redhat-marketplace-scs7t\" (UID: \"4990ec88-046b-4039-8ae8-062cf8c97b14\") " pod="openshift-marketplace/redhat-marketplace-scs7t" Nov 29 05:15:13 crc kubenswrapper[4799]: I1129 05:15:13.953884 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxmqp\" (UniqueName: \"kubernetes.io/projected/4990ec88-046b-4039-8ae8-062cf8c97b14-kube-api-access-mxmqp\") pod \"redhat-marketplace-scs7t\" (UID: \"4990ec88-046b-4039-8ae8-062cf8c97b14\") " pod="openshift-marketplace/redhat-marketplace-scs7t" Nov 29 05:15:14 crc kubenswrapper[4799]: I1129 05:15:14.057194 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4990ec88-046b-4039-8ae8-062cf8c97b14-catalog-content\") pod \"redhat-marketplace-scs7t\" (UID: \"4990ec88-046b-4039-8ae8-062cf8c97b14\") " pod="openshift-marketplace/redhat-marketplace-scs7t" Nov 29 05:15:14 crc kubenswrapper[4799]: I1129 05:15:14.057462 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4990ec88-046b-4039-8ae8-062cf8c97b14-utilities\") pod \"redhat-marketplace-scs7t\" (UID: \"4990ec88-046b-4039-8ae8-062cf8c97b14\") " pod="openshift-marketplace/redhat-marketplace-scs7t" Nov 29 05:15:14 crc kubenswrapper[4799]: I1129 05:15:14.057537 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxmqp\" (UniqueName: \"kubernetes.io/projected/4990ec88-046b-4039-8ae8-062cf8c97b14-kube-api-access-mxmqp\") pod \"redhat-marketplace-scs7t\" (UID: \"4990ec88-046b-4039-8ae8-062cf8c97b14\") " pod="openshift-marketplace/redhat-marketplace-scs7t" Nov 29 05:15:14 crc kubenswrapper[4799]: I1129 05:15:14.058588 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4990ec88-046b-4039-8ae8-062cf8c97b14-catalog-content\") pod \"redhat-marketplace-scs7t\" (UID: \"4990ec88-046b-4039-8ae8-062cf8c97b14\") " pod="openshift-marketplace/redhat-marketplace-scs7t" Nov 29 05:15:14 crc kubenswrapper[4799]: I1129 05:15:14.058773 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4990ec88-046b-4039-8ae8-062cf8c97b14-utilities\") pod \"redhat-marketplace-scs7t\" (UID: \"4990ec88-046b-4039-8ae8-062cf8c97b14\") " pod="openshift-marketplace/redhat-marketplace-scs7t" Nov 29 05:15:14 crc kubenswrapper[4799]: I1129 05:15:14.088062 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxmqp\" (UniqueName: \"kubernetes.io/projected/4990ec88-046b-4039-8ae8-062cf8c97b14-kube-api-access-mxmqp\") pod \"redhat-marketplace-scs7t\" (UID: \"4990ec88-046b-4039-8ae8-062cf8c97b14\") " pod="openshift-marketplace/redhat-marketplace-scs7t" Nov 29 05:15:14 crc kubenswrapper[4799]: I1129 05:15:14.167422 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-scs7t" Nov 29 05:15:14 crc kubenswrapper[4799]: I1129 05:15:14.716280 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-scs7t"] Nov 29 05:15:15 crc kubenswrapper[4799]: I1129 05:15:15.522080 4799 generic.go:334] "Generic (PLEG): container finished" podID="4990ec88-046b-4039-8ae8-062cf8c97b14" containerID="cd9871faa5982e5eff2e51db535f2ebdc532507ace8d3dd717a3a184f62fc74c" exitCode=0 Nov 29 05:15:15 crc kubenswrapper[4799]: I1129 05:15:15.522172 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-scs7t" event={"ID":"4990ec88-046b-4039-8ae8-062cf8c97b14","Type":"ContainerDied","Data":"cd9871faa5982e5eff2e51db535f2ebdc532507ace8d3dd717a3a184f62fc74c"} Nov 29 05:15:15 crc kubenswrapper[4799]: I1129 05:15:15.522865 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-scs7t" event={"ID":"4990ec88-046b-4039-8ae8-062cf8c97b14","Type":"ContainerStarted","Data":"9ddaa6f4a4495f07273e76099ee1ae7d3db63251f8a8bad2244c0f3632727e73"} Nov 29 05:15:16 crc kubenswrapper[4799]: I1129 05:15:16.543824 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-scs7t" event={"ID":"4990ec88-046b-4039-8ae8-062cf8c97b14","Type":"ContainerStarted","Data":"6677157479a521e4b3291e1a91bcc638f0be35dac97e2a6d378fa0c4b001af19"} Nov 29 05:15:16 crc kubenswrapper[4799]: I1129 05:15:16.547427 4799 generic.go:334] "Generic (PLEG): container finished" podID="92efc83b-d2e4-44f7-8ede-db45686a6c0f" containerID="059d74fc5ab3f32e2c9c104b552f40c0b550335427e8ac36026515c8854b5bfd" exitCode=0 Nov 29 05:15:16 crc kubenswrapper[4799]: I1129 05:15:16.547527 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6rm8k" event={"ID":"92efc83b-d2e4-44f7-8ede-db45686a6c0f","Type":"ContainerDied","Data":"059d74fc5ab3f32e2c9c104b552f40c0b550335427e8ac36026515c8854b5bfd"} Nov 29 05:15:17 crc kubenswrapper[4799]: I1129 05:15:17.566431 4799 generic.go:334] "Generic (PLEG): container finished" podID="4990ec88-046b-4039-8ae8-062cf8c97b14" containerID="6677157479a521e4b3291e1a91bcc638f0be35dac97e2a6d378fa0c4b001af19" exitCode=0 Nov 29 05:15:17 crc kubenswrapper[4799]: I1129 05:15:17.568711 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-scs7t" event={"ID":"4990ec88-046b-4039-8ae8-062cf8c97b14","Type":"ContainerDied","Data":"6677157479a521e4b3291e1a91bcc638f0be35dac97e2a6d378fa0c4b001af19"} Nov 29 05:15:17 crc kubenswrapper[4799]: I1129 05:15:17.568757 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-scs7t" event={"ID":"4990ec88-046b-4039-8ae8-062cf8c97b14","Type":"ContainerStarted","Data":"fd46a3d5c6f27e47ff9f2cb999efeb60def13c6f9183eb4ff29df9a6289ef5af"} Nov 29 05:15:17 crc kubenswrapper[4799]: I1129 05:15:17.597880 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-scs7t" podStartSLOduration=3.0784034829999998 podStartE2EDuration="4.597853303s" podCreationTimestamp="2025-11-29 05:15:13 +0000 UTC" firstStartedPulling="2025-11-29 05:15:15.52467143 +0000 UTC m=+2191.167601840" lastFinishedPulling="2025-11-29 05:15:17.04412122 +0000 UTC m=+2192.687051660" observedRunningTime="2025-11-29 05:15:17.58749264 +0000 UTC m=+2193.230423030" watchObservedRunningTime="2025-11-29 05:15:17.597853303 +0000 UTC m=+2193.240783713" Nov 29 05:15:18 crc kubenswrapper[4799]: I1129 05:15:18.096359 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6rm8k" Nov 29 05:15:18 crc kubenswrapper[4799]: I1129 05:15:18.161361 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/92efc83b-d2e4-44f7-8ede-db45686a6c0f-inventory\") pod \"92efc83b-d2e4-44f7-8ede-db45686a6c0f\" (UID: \"92efc83b-d2e4-44f7-8ede-db45686a6c0f\") " Nov 29 05:15:18 crc kubenswrapper[4799]: I1129 05:15:18.161721 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/92efc83b-d2e4-44f7-8ede-db45686a6c0f-ceph\") pod \"92efc83b-d2e4-44f7-8ede-db45686a6c0f\" (UID: \"92efc83b-d2e4-44f7-8ede-db45686a6c0f\") " Nov 29 05:15:18 crc kubenswrapper[4799]: I1129 05:15:18.161876 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/92efc83b-d2e4-44f7-8ede-db45686a6c0f-ssh-key\") pod \"92efc83b-d2e4-44f7-8ede-db45686a6c0f\" (UID: \"92efc83b-d2e4-44f7-8ede-db45686a6c0f\") " Nov 29 05:15:18 crc kubenswrapper[4799]: I1129 05:15:18.161981 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7kntp\" (UniqueName: \"kubernetes.io/projected/92efc83b-d2e4-44f7-8ede-db45686a6c0f-kube-api-access-7kntp\") pod \"92efc83b-d2e4-44f7-8ede-db45686a6c0f\" (UID: \"92efc83b-d2e4-44f7-8ede-db45686a6c0f\") " Nov 29 05:15:18 crc kubenswrapper[4799]: I1129 05:15:18.170132 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92efc83b-d2e4-44f7-8ede-db45686a6c0f-ceph" (OuterVolumeSpecName: "ceph") pod "92efc83b-d2e4-44f7-8ede-db45686a6c0f" (UID: "92efc83b-d2e4-44f7-8ede-db45686a6c0f"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:15:18 crc kubenswrapper[4799]: I1129 05:15:18.177323 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92efc83b-d2e4-44f7-8ede-db45686a6c0f-kube-api-access-7kntp" (OuterVolumeSpecName: "kube-api-access-7kntp") pod "92efc83b-d2e4-44f7-8ede-db45686a6c0f" (UID: "92efc83b-d2e4-44f7-8ede-db45686a6c0f"). InnerVolumeSpecName "kube-api-access-7kntp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:15:18 crc kubenswrapper[4799]: I1129 05:15:18.199572 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92efc83b-d2e4-44f7-8ede-db45686a6c0f-inventory" (OuterVolumeSpecName: "inventory") pod "92efc83b-d2e4-44f7-8ede-db45686a6c0f" (UID: "92efc83b-d2e4-44f7-8ede-db45686a6c0f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:15:18 crc kubenswrapper[4799]: I1129 05:15:18.204104 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92efc83b-d2e4-44f7-8ede-db45686a6c0f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "92efc83b-d2e4-44f7-8ede-db45686a6c0f" (UID: "92efc83b-d2e4-44f7-8ede-db45686a6c0f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:15:18 crc kubenswrapper[4799]: I1129 05:15:18.265020 4799 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/92efc83b-d2e4-44f7-8ede-db45686a6c0f-ceph\") on node \"crc\" DevicePath \"\"" Nov 29 05:15:18 crc kubenswrapper[4799]: I1129 05:15:18.265060 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/92efc83b-d2e4-44f7-8ede-db45686a6c0f-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 29 05:15:18 crc kubenswrapper[4799]: I1129 05:15:18.265074 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7kntp\" (UniqueName: \"kubernetes.io/projected/92efc83b-d2e4-44f7-8ede-db45686a6c0f-kube-api-access-7kntp\") on node \"crc\" DevicePath \"\"" Nov 29 05:15:18 crc kubenswrapper[4799]: I1129 05:15:18.265091 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/92efc83b-d2e4-44f7-8ede-db45686a6c0f-inventory\") on node \"crc\" DevicePath \"\"" Nov 29 05:15:18 crc kubenswrapper[4799]: I1129 05:15:18.579931 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6rm8k" event={"ID":"92efc83b-d2e4-44f7-8ede-db45686a6c0f","Type":"ContainerDied","Data":"2c1ca7ce78c1ba2d982ef5934b9d7a7be6d23a7b11a68cb72e748fad442cc516"} Nov 29 05:15:18 crc kubenswrapper[4799]: I1129 05:15:18.580033 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2c1ca7ce78c1ba2d982ef5934b9d7a7be6d23a7b11a68cb72e748fad442cc516" Nov 29 05:15:18 crc kubenswrapper[4799]: I1129 05:15:18.579960 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6rm8k" Nov 29 05:15:18 crc kubenswrapper[4799]: I1129 05:15:18.686087 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k7vvz"] Nov 29 05:15:18 crc kubenswrapper[4799]: E1129 05:15:18.686549 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92efc83b-d2e4-44f7-8ede-db45686a6c0f" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 29 05:15:18 crc kubenswrapper[4799]: I1129 05:15:18.686573 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="92efc83b-d2e4-44f7-8ede-db45686a6c0f" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 29 05:15:18 crc kubenswrapper[4799]: I1129 05:15:18.686850 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="92efc83b-d2e4-44f7-8ede-db45686a6c0f" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 29 05:15:18 crc kubenswrapper[4799]: I1129 05:15:18.687890 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k7vvz" Nov 29 05:15:18 crc kubenswrapper[4799]: I1129 05:15:18.690816 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 29 05:15:18 crc kubenswrapper[4799]: I1129 05:15:18.691088 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 29 05:15:18 crc kubenswrapper[4799]: I1129 05:15:18.691182 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ljwtk" Nov 29 05:15:18 crc kubenswrapper[4799]: I1129 05:15:18.695769 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 29 05:15:18 crc kubenswrapper[4799]: I1129 05:15:18.695856 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 29 05:15:18 crc kubenswrapper[4799]: I1129 05:15:18.699707 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k7vvz"] Nov 29 05:15:18 crc kubenswrapper[4799]: I1129 05:15:18.775931 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/56293cbb-ed95-4541-b322-8e86233a8ff5-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-k7vvz\" (UID: \"56293cbb-ed95-4541-b322-8e86233a8ff5\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k7vvz" Nov 29 05:15:18 crc kubenswrapper[4799]: I1129 05:15:18.776042 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/56293cbb-ed95-4541-b322-8e86233a8ff5-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-k7vvz\" (UID: \"56293cbb-ed95-4541-b322-8e86233a8ff5\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k7vvz" Nov 29 05:15:18 crc kubenswrapper[4799]: I1129 05:15:18.776106 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmzfx\" (UniqueName: \"kubernetes.io/projected/56293cbb-ed95-4541-b322-8e86233a8ff5-kube-api-access-lmzfx\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-k7vvz\" (UID: \"56293cbb-ed95-4541-b322-8e86233a8ff5\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k7vvz" Nov 29 05:15:18 crc kubenswrapper[4799]: I1129 05:15:18.776159 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/56293cbb-ed95-4541-b322-8e86233a8ff5-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-k7vvz\" (UID: \"56293cbb-ed95-4541-b322-8e86233a8ff5\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k7vvz" Nov 29 05:15:18 crc kubenswrapper[4799]: I1129 05:15:18.877220 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/56293cbb-ed95-4541-b322-8e86233a8ff5-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-k7vvz\" (UID: \"56293cbb-ed95-4541-b322-8e86233a8ff5\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k7vvz" Nov 29 05:15:18 crc kubenswrapper[4799]: I1129 05:15:18.877353 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/56293cbb-ed95-4541-b322-8e86233a8ff5-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-k7vvz\" (UID: \"56293cbb-ed95-4541-b322-8e86233a8ff5\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k7vvz" Nov 29 05:15:18 crc kubenswrapper[4799]: I1129 05:15:18.877400 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/56293cbb-ed95-4541-b322-8e86233a8ff5-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-k7vvz\" (UID: \"56293cbb-ed95-4541-b322-8e86233a8ff5\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k7vvz" Nov 29 05:15:18 crc kubenswrapper[4799]: I1129 05:15:18.877428 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmzfx\" (UniqueName: \"kubernetes.io/projected/56293cbb-ed95-4541-b322-8e86233a8ff5-kube-api-access-lmzfx\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-k7vvz\" (UID: \"56293cbb-ed95-4541-b322-8e86233a8ff5\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k7vvz" Nov 29 05:15:18 crc kubenswrapper[4799]: I1129 05:15:18.883823 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/56293cbb-ed95-4541-b322-8e86233a8ff5-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-k7vvz\" (UID: \"56293cbb-ed95-4541-b322-8e86233a8ff5\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k7vvz" Nov 29 05:15:18 crc kubenswrapper[4799]: I1129 05:15:18.883902 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/56293cbb-ed95-4541-b322-8e86233a8ff5-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-k7vvz\" (UID: \"56293cbb-ed95-4541-b322-8e86233a8ff5\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k7vvz" Nov 29 05:15:18 crc kubenswrapper[4799]: I1129 05:15:18.899598 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/56293cbb-ed95-4541-b322-8e86233a8ff5-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-k7vvz\" (UID: \"56293cbb-ed95-4541-b322-8e86233a8ff5\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k7vvz" Nov 29 05:15:18 crc kubenswrapper[4799]: I1129 05:15:18.903437 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmzfx\" (UniqueName: \"kubernetes.io/projected/56293cbb-ed95-4541-b322-8e86233a8ff5-kube-api-access-lmzfx\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-k7vvz\" (UID: \"56293cbb-ed95-4541-b322-8e86233a8ff5\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k7vvz" Nov 29 05:15:19 crc kubenswrapper[4799]: I1129 05:15:19.007905 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k7vvz" Nov 29 05:15:19 crc kubenswrapper[4799]: I1129 05:15:19.590412 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k7vvz"] Nov 29 05:15:20 crc kubenswrapper[4799]: I1129 05:15:20.608385 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k7vvz" event={"ID":"56293cbb-ed95-4541-b322-8e86233a8ff5","Type":"ContainerStarted","Data":"dc1a2534938c379d66f94d4b1c742fdeae40b50c8dbe79554885dbeb19a77bcd"} Nov 29 05:15:20 crc kubenswrapper[4799]: I1129 05:15:20.609461 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k7vvz" event={"ID":"56293cbb-ed95-4541-b322-8e86233a8ff5","Type":"ContainerStarted","Data":"0dfda104a06857a2e43bca928edeb30be3ce54251b9d268571180c1934b91d08"} Nov 29 05:15:20 crc kubenswrapper[4799]: I1129 05:15:20.633268 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k7vvz" podStartSLOduration=2.152777316 podStartE2EDuration="2.633243176s" podCreationTimestamp="2025-11-29 05:15:18 +0000 UTC" firstStartedPulling="2025-11-29 05:15:19.600473297 +0000 UTC m=+2195.243403697" lastFinishedPulling="2025-11-29 05:15:20.080939157 +0000 UTC m=+2195.723869557" observedRunningTime="2025-11-29 05:15:20.626409998 +0000 UTC m=+2196.269340408" watchObservedRunningTime="2025-11-29 05:15:20.633243176 +0000 UTC m=+2196.276173586" Nov 29 05:15:24 crc kubenswrapper[4799]: I1129 05:15:24.170293 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-scs7t" Nov 29 05:15:24 crc kubenswrapper[4799]: I1129 05:15:24.170868 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-scs7t" Nov 29 05:15:24 crc kubenswrapper[4799]: I1129 05:15:24.258260 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-scs7t" Nov 29 05:15:24 crc kubenswrapper[4799]: I1129 05:15:24.750811 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-scs7t" Nov 29 05:15:24 crc kubenswrapper[4799]: I1129 05:15:24.819170 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-scs7t"] Nov 29 05:15:26 crc kubenswrapper[4799]: I1129 05:15:26.683912 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-scs7t" podUID="4990ec88-046b-4039-8ae8-062cf8c97b14" containerName="registry-server" containerID="cri-o://fd46a3d5c6f27e47ff9f2cb999efeb60def13c6f9183eb4ff29df9a6289ef5af" gracePeriod=2 Nov 29 05:15:27 crc kubenswrapper[4799]: I1129 05:15:27.244393 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-scs7t" Nov 29 05:15:27 crc kubenswrapper[4799]: I1129 05:15:27.302287 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4990ec88-046b-4039-8ae8-062cf8c97b14-utilities\") pod \"4990ec88-046b-4039-8ae8-062cf8c97b14\" (UID: \"4990ec88-046b-4039-8ae8-062cf8c97b14\") " Nov 29 05:15:27 crc kubenswrapper[4799]: I1129 05:15:27.302846 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4990ec88-046b-4039-8ae8-062cf8c97b14-catalog-content\") pod \"4990ec88-046b-4039-8ae8-062cf8c97b14\" (UID: \"4990ec88-046b-4039-8ae8-062cf8c97b14\") " Nov 29 05:15:27 crc kubenswrapper[4799]: I1129 05:15:27.302922 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mxmqp\" (UniqueName: \"kubernetes.io/projected/4990ec88-046b-4039-8ae8-062cf8c97b14-kube-api-access-mxmqp\") pod \"4990ec88-046b-4039-8ae8-062cf8c97b14\" (UID: \"4990ec88-046b-4039-8ae8-062cf8c97b14\") " Nov 29 05:15:27 crc kubenswrapper[4799]: I1129 05:15:27.303907 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4990ec88-046b-4039-8ae8-062cf8c97b14-utilities" (OuterVolumeSpecName: "utilities") pod "4990ec88-046b-4039-8ae8-062cf8c97b14" (UID: "4990ec88-046b-4039-8ae8-062cf8c97b14"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:15:27 crc kubenswrapper[4799]: I1129 05:15:27.323603 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4990ec88-046b-4039-8ae8-062cf8c97b14-kube-api-access-mxmqp" (OuterVolumeSpecName: "kube-api-access-mxmqp") pod "4990ec88-046b-4039-8ae8-062cf8c97b14" (UID: "4990ec88-046b-4039-8ae8-062cf8c97b14"). InnerVolumeSpecName "kube-api-access-mxmqp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:15:27 crc kubenswrapper[4799]: I1129 05:15:27.324006 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4990ec88-046b-4039-8ae8-062cf8c97b14-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4990ec88-046b-4039-8ae8-062cf8c97b14" (UID: "4990ec88-046b-4039-8ae8-062cf8c97b14"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:15:27 crc kubenswrapper[4799]: I1129 05:15:27.405849 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4990ec88-046b-4039-8ae8-062cf8c97b14-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 05:15:27 crc kubenswrapper[4799]: I1129 05:15:27.405919 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mxmqp\" (UniqueName: \"kubernetes.io/projected/4990ec88-046b-4039-8ae8-062cf8c97b14-kube-api-access-mxmqp\") on node \"crc\" DevicePath \"\"" Nov 29 05:15:27 crc kubenswrapper[4799]: I1129 05:15:27.405942 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4990ec88-046b-4039-8ae8-062cf8c97b14-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 05:15:27 crc kubenswrapper[4799]: I1129 05:15:27.701292 4799 generic.go:334] "Generic (PLEG): container finished" podID="4990ec88-046b-4039-8ae8-062cf8c97b14" containerID="fd46a3d5c6f27e47ff9f2cb999efeb60def13c6f9183eb4ff29df9a6289ef5af" exitCode=0 Nov 29 05:15:27 crc kubenswrapper[4799]: I1129 05:15:27.701384 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-scs7t" event={"ID":"4990ec88-046b-4039-8ae8-062cf8c97b14","Type":"ContainerDied","Data":"fd46a3d5c6f27e47ff9f2cb999efeb60def13c6f9183eb4ff29df9a6289ef5af"} Nov 29 05:15:27 crc kubenswrapper[4799]: I1129 05:15:27.701415 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-scs7t" Nov 29 05:15:27 crc kubenswrapper[4799]: I1129 05:15:27.701477 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-scs7t" event={"ID":"4990ec88-046b-4039-8ae8-062cf8c97b14","Type":"ContainerDied","Data":"9ddaa6f4a4495f07273e76099ee1ae7d3db63251f8a8bad2244c0f3632727e73"} Nov 29 05:15:27 crc kubenswrapper[4799]: I1129 05:15:27.701535 4799 scope.go:117] "RemoveContainer" containerID="fd46a3d5c6f27e47ff9f2cb999efeb60def13c6f9183eb4ff29df9a6289ef5af" Nov 29 05:15:27 crc kubenswrapper[4799]: I1129 05:15:27.764103 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-scs7t"] Nov 29 05:15:27 crc kubenswrapper[4799]: I1129 05:15:27.765356 4799 scope.go:117] "RemoveContainer" containerID="6677157479a521e4b3291e1a91bcc638f0be35dac97e2a6d378fa0c4b001af19" Nov 29 05:15:27 crc kubenswrapper[4799]: I1129 05:15:27.781055 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-scs7t"] Nov 29 05:15:27 crc kubenswrapper[4799]: I1129 05:15:27.803180 4799 scope.go:117] "RemoveContainer" containerID="cd9871faa5982e5eff2e51db535f2ebdc532507ace8d3dd717a3a184f62fc74c" Nov 29 05:15:27 crc kubenswrapper[4799]: I1129 05:15:27.870213 4799 scope.go:117] "RemoveContainer" containerID="fd46a3d5c6f27e47ff9f2cb999efeb60def13c6f9183eb4ff29df9a6289ef5af" Nov 29 05:15:27 crc kubenswrapper[4799]: E1129 05:15:27.870954 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd46a3d5c6f27e47ff9f2cb999efeb60def13c6f9183eb4ff29df9a6289ef5af\": container with ID starting with fd46a3d5c6f27e47ff9f2cb999efeb60def13c6f9183eb4ff29df9a6289ef5af not found: ID does not exist" containerID="fd46a3d5c6f27e47ff9f2cb999efeb60def13c6f9183eb4ff29df9a6289ef5af" Nov 29 05:15:27 crc kubenswrapper[4799]: I1129 05:15:27.871014 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd46a3d5c6f27e47ff9f2cb999efeb60def13c6f9183eb4ff29df9a6289ef5af"} err="failed to get container status \"fd46a3d5c6f27e47ff9f2cb999efeb60def13c6f9183eb4ff29df9a6289ef5af\": rpc error: code = NotFound desc = could not find container \"fd46a3d5c6f27e47ff9f2cb999efeb60def13c6f9183eb4ff29df9a6289ef5af\": container with ID starting with fd46a3d5c6f27e47ff9f2cb999efeb60def13c6f9183eb4ff29df9a6289ef5af not found: ID does not exist" Nov 29 05:15:27 crc kubenswrapper[4799]: I1129 05:15:27.871107 4799 scope.go:117] "RemoveContainer" containerID="6677157479a521e4b3291e1a91bcc638f0be35dac97e2a6d378fa0c4b001af19" Nov 29 05:15:27 crc kubenswrapper[4799]: E1129 05:15:27.871597 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6677157479a521e4b3291e1a91bcc638f0be35dac97e2a6d378fa0c4b001af19\": container with ID starting with 6677157479a521e4b3291e1a91bcc638f0be35dac97e2a6d378fa0c4b001af19 not found: ID does not exist" containerID="6677157479a521e4b3291e1a91bcc638f0be35dac97e2a6d378fa0c4b001af19" Nov 29 05:15:27 crc kubenswrapper[4799]: I1129 05:15:27.871671 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6677157479a521e4b3291e1a91bcc638f0be35dac97e2a6d378fa0c4b001af19"} err="failed to get container status \"6677157479a521e4b3291e1a91bcc638f0be35dac97e2a6d378fa0c4b001af19\": rpc error: code = NotFound desc = could not find container \"6677157479a521e4b3291e1a91bcc638f0be35dac97e2a6d378fa0c4b001af19\": container with ID starting with 6677157479a521e4b3291e1a91bcc638f0be35dac97e2a6d378fa0c4b001af19 not found: ID does not exist" Nov 29 05:15:27 crc kubenswrapper[4799]: I1129 05:15:27.871723 4799 scope.go:117] "RemoveContainer" containerID="cd9871faa5982e5eff2e51db535f2ebdc532507ace8d3dd717a3a184f62fc74c" Nov 29 05:15:27 crc kubenswrapper[4799]: E1129 05:15:27.872370 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd9871faa5982e5eff2e51db535f2ebdc532507ace8d3dd717a3a184f62fc74c\": container with ID starting with cd9871faa5982e5eff2e51db535f2ebdc532507ace8d3dd717a3a184f62fc74c not found: ID does not exist" containerID="cd9871faa5982e5eff2e51db535f2ebdc532507ace8d3dd717a3a184f62fc74c" Nov 29 05:15:27 crc kubenswrapper[4799]: I1129 05:15:27.872428 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd9871faa5982e5eff2e51db535f2ebdc532507ace8d3dd717a3a184f62fc74c"} err="failed to get container status \"cd9871faa5982e5eff2e51db535f2ebdc532507ace8d3dd717a3a184f62fc74c\": rpc error: code = NotFound desc = could not find container \"cd9871faa5982e5eff2e51db535f2ebdc532507ace8d3dd717a3a184f62fc74c\": container with ID starting with cd9871faa5982e5eff2e51db535f2ebdc532507ace8d3dd717a3a184f62fc74c not found: ID does not exist" Nov 29 05:15:28 crc kubenswrapper[4799]: I1129 05:15:28.679706 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4990ec88-046b-4039-8ae8-062cf8c97b14" path="/var/lib/kubelet/pods/4990ec88-046b-4039-8ae8-062cf8c97b14/volumes" Nov 29 05:15:37 crc kubenswrapper[4799]: I1129 05:15:37.497634 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 05:15:37 crc kubenswrapper[4799]: I1129 05:15:37.498832 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 05:15:53 crc kubenswrapper[4799]: I1129 05:15:53.034949 4799 scope.go:117] "RemoveContainer" containerID="4cdf55e51f971d2c8da1cd6d4fe65f3ad6e77e9db99c9d9f0bb9edbd026a4d63" Nov 29 05:16:07 crc kubenswrapper[4799]: I1129 05:16:07.497713 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 05:16:07 crc kubenswrapper[4799]: I1129 05:16:07.498670 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 05:16:07 crc kubenswrapper[4799]: I1129 05:16:07.498757 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-26t88" Nov 29 05:16:07 crc kubenswrapper[4799]: I1129 05:16:07.500075 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d0281928bf1426881365580d5726280995a377a80c626b0654de38a8828878e1"} pod="openshift-machine-config-operator/machine-config-daemon-26t88" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 29 05:16:07 crc kubenswrapper[4799]: I1129 05:16:07.500185 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" containerID="cri-o://d0281928bf1426881365580d5726280995a377a80c626b0654de38a8828878e1" gracePeriod=600 Nov 29 05:16:07 crc kubenswrapper[4799]: E1129 05:16:07.628236 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:16:08 crc kubenswrapper[4799]: I1129 05:16:08.215139 4799 generic.go:334] "Generic (PLEG): container finished" podID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerID="d0281928bf1426881365580d5726280995a377a80c626b0654de38a8828878e1" exitCode=0 Nov 29 05:16:08 crc kubenswrapper[4799]: I1129 05:16:08.215341 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" event={"ID":"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8","Type":"ContainerDied","Data":"d0281928bf1426881365580d5726280995a377a80c626b0654de38a8828878e1"} Nov 29 05:16:08 crc kubenswrapper[4799]: I1129 05:16:08.215525 4799 scope.go:117] "RemoveContainer" containerID="4a35df0e0c53bb8deb879dbc720f6c155860ecc2180303a28b55f87c720cacd9" Nov 29 05:16:08 crc kubenswrapper[4799]: I1129 05:16:08.217135 4799 scope.go:117] "RemoveContainer" containerID="d0281928bf1426881365580d5726280995a377a80c626b0654de38a8828878e1" Nov 29 05:16:08 crc kubenswrapper[4799]: E1129 05:16:08.217745 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:16:10 crc kubenswrapper[4799]: I1129 05:16:10.239864 4799 generic.go:334] "Generic (PLEG): container finished" podID="56293cbb-ed95-4541-b322-8e86233a8ff5" containerID="dc1a2534938c379d66f94d4b1c742fdeae40b50c8dbe79554885dbeb19a77bcd" exitCode=0 Nov 29 05:16:10 crc kubenswrapper[4799]: I1129 05:16:10.240023 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k7vvz" event={"ID":"56293cbb-ed95-4541-b322-8e86233a8ff5","Type":"ContainerDied","Data":"dc1a2534938c379d66f94d4b1c742fdeae40b50c8dbe79554885dbeb19a77bcd"} Nov 29 05:16:11 crc kubenswrapper[4799]: I1129 05:16:11.756060 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k7vvz" Nov 29 05:16:11 crc kubenswrapper[4799]: I1129 05:16:11.930882 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/56293cbb-ed95-4541-b322-8e86233a8ff5-ssh-key\") pod \"56293cbb-ed95-4541-b322-8e86233a8ff5\" (UID: \"56293cbb-ed95-4541-b322-8e86233a8ff5\") " Nov 29 05:16:11 crc kubenswrapper[4799]: I1129 05:16:11.930947 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/56293cbb-ed95-4541-b322-8e86233a8ff5-ceph\") pod \"56293cbb-ed95-4541-b322-8e86233a8ff5\" (UID: \"56293cbb-ed95-4541-b322-8e86233a8ff5\") " Nov 29 05:16:11 crc kubenswrapper[4799]: I1129 05:16:11.930983 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/56293cbb-ed95-4541-b322-8e86233a8ff5-inventory\") pod \"56293cbb-ed95-4541-b322-8e86233a8ff5\" (UID: \"56293cbb-ed95-4541-b322-8e86233a8ff5\") " Nov 29 05:16:11 crc kubenswrapper[4799]: I1129 05:16:11.931011 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lmzfx\" (UniqueName: \"kubernetes.io/projected/56293cbb-ed95-4541-b322-8e86233a8ff5-kube-api-access-lmzfx\") pod \"56293cbb-ed95-4541-b322-8e86233a8ff5\" (UID: \"56293cbb-ed95-4541-b322-8e86233a8ff5\") " Nov 29 05:16:11 crc kubenswrapper[4799]: I1129 05:16:11.946930 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56293cbb-ed95-4541-b322-8e86233a8ff5-kube-api-access-lmzfx" (OuterVolumeSpecName: "kube-api-access-lmzfx") pod "56293cbb-ed95-4541-b322-8e86233a8ff5" (UID: "56293cbb-ed95-4541-b322-8e86233a8ff5"). InnerVolumeSpecName "kube-api-access-lmzfx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:16:11 crc kubenswrapper[4799]: I1129 05:16:11.947653 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56293cbb-ed95-4541-b322-8e86233a8ff5-ceph" (OuterVolumeSpecName: "ceph") pod "56293cbb-ed95-4541-b322-8e86233a8ff5" (UID: "56293cbb-ed95-4541-b322-8e86233a8ff5"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:16:11 crc kubenswrapper[4799]: I1129 05:16:11.966929 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56293cbb-ed95-4541-b322-8e86233a8ff5-inventory" (OuterVolumeSpecName: "inventory") pod "56293cbb-ed95-4541-b322-8e86233a8ff5" (UID: "56293cbb-ed95-4541-b322-8e86233a8ff5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:16:11 crc kubenswrapper[4799]: I1129 05:16:11.968008 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56293cbb-ed95-4541-b322-8e86233a8ff5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "56293cbb-ed95-4541-b322-8e86233a8ff5" (UID: "56293cbb-ed95-4541-b322-8e86233a8ff5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:16:12 crc kubenswrapper[4799]: I1129 05:16:12.034322 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/56293cbb-ed95-4541-b322-8e86233a8ff5-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 29 05:16:12 crc kubenswrapper[4799]: I1129 05:16:12.034378 4799 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/56293cbb-ed95-4541-b322-8e86233a8ff5-ceph\") on node \"crc\" DevicePath \"\"" Nov 29 05:16:12 crc kubenswrapper[4799]: I1129 05:16:12.034390 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/56293cbb-ed95-4541-b322-8e86233a8ff5-inventory\") on node \"crc\" DevicePath \"\"" Nov 29 05:16:12 crc kubenswrapper[4799]: I1129 05:16:12.034404 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lmzfx\" (UniqueName: \"kubernetes.io/projected/56293cbb-ed95-4541-b322-8e86233a8ff5-kube-api-access-lmzfx\") on node \"crc\" DevicePath \"\"" Nov 29 05:16:12 crc kubenswrapper[4799]: I1129 05:16:12.265907 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k7vvz" event={"ID":"56293cbb-ed95-4541-b322-8e86233a8ff5","Type":"ContainerDied","Data":"0dfda104a06857a2e43bca928edeb30be3ce54251b9d268571180c1934b91d08"} Nov 29 05:16:12 crc kubenswrapper[4799]: I1129 05:16:12.265973 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k7vvz" Nov 29 05:16:12 crc kubenswrapper[4799]: I1129 05:16:12.265978 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0dfda104a06857a2e43bca928edeb30be3ce54251b9d268571180c1934b91d08" Nov 29 05:16:12 crc kubenswrapper[4799]: I1129 05:16:12.408126 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-9hpsr"] Nov 29 05:16:12 crc kubenswrapper[4799]: E1129 05:16:12.408775 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56293cbb-ed95-4541-b322-8e86233a8ff5" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 29 05:16:12 crc kubenswrapper[4799]: I1129 05:16:12.408833 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="56293cbb-ed95-4541-b322-8e86233a8ff5" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 29 05:16:12 crc kubenswrapper[4799]: E1129 05:16:12.408870 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4990ec88-046b-4039-8ae8-062cf8c97b14" containerName="extract-utilities" Nov 29 05:16:12 crc kubenswrapper[4799]: I1129 05:16:12.408878 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="4990ec88-046b-4039-8ae8-062cf8c97b14" containerName="extract-utilities" Nov 29 05:16:12 crc kubenswrapper[4799]: E1129 05:16:12.408893 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4990ec88-046b-4039-8ae8-062cf8c97b14" containerName="registry-server" Nov 29 05:16:12 crc kubenswrapper[4799]: I1129 05:16:12.408902 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="4990ec88-046b-4039-8ae8-062cf8c97b14" containerName="registry-server" Nov 29 05:16:12 crc kubenswrapper[4799]: E1129 05:16:12.408954 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4990ec88-046b-4039-8ae8-062cf8c97b14" containerName="extract-content" Nov 29 05:16:12 crc kubenswrapper[4799]: I1129 05:16:12.408980 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="4990ec88-046b-4039-8ae8-062cf8c97b14" containerName="extract-content" Nov 29 05:16:12 crc kubenswrapper[4799]: I1129 05:16:12.409241 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="4990ec88-046b-4039-8ae8-062cf8c97b14" containerName="registry-server" Nov 29 05:16:12 crc kubenswrapper[4799]: I1129 05:16:12.409269 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="56293cbb-ed95-4541-b322-8e86233a8ff5" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 29 05:16:12 crc kubenswrapper[4799]: I1129 05:16:12.410298 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-9hpsr" Nov 29 05:16:12 crc kubenswrapper[4799]: I1129 05:16:12.413344 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 29 05:16:12 crc kubenswrapper[4799]: I1129 05:16:12.417617 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 29 05:16:12 crc kubenswrapper[4799]: I1129 05:16:12.418086 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ljwtk" Nov 29 05:16:12 crc kubenswrapper[4799]: I1129 05:16:12.418382 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 29 05:16:12 crc kubenswrapper[4799]: I1129 05:16:12.418611 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 29 05:16:12 crc kubenswrapper[4799]: I1129 05:16:12.422227 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-9hpsr"] Nov 29 05:16:12 crc kubenswrapper[4799]: I1129 05:16:12.550059 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xx2st\" (UniqueName: \"kubernetes.io/projected/3a84e2da-773c-4bf7-8f99-a84276791e78-kube-api-access-xx2st\") pod \"ssh-known-hosts-edpm-deployment-9hpsr\" (UID: \"3a84e2da-773c-4bf7-8f99-a84276791e78\") " pod="openstack/ssh-known-hosts-edpm-deployment-9hpsr" Nov 29 05:16:12 crc kubenswrapper[4799]: I1129 05:16:12.550143 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3a84e2da-773c-4bf7-8f99-a84276791e78-ceph\") pod \"ssh-known-hosts-edpm-deployment-9hpsr\" (UID: \"3a84e2da-773c-4bf7-8f99-a84276791e78\") " pod="openstack/ssh-known-hosts-edpm-deployment-9hpsr" Nov 29 05:16:12 crc kubenswrapper[4799]: I1129 05:16:12.550451 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/3a84e2da-773c-4bf7-8f99-a84276791e78-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-9hpsr\" (UID: \"3a84e2da-773c-4bf7-8f99-a84276791e78\") " pod="openstack/ssh-known-hosts-edpm-deployment-9hpsr" Nov 29 05:16:12 crc kubenswrapper[4799]: I1129 05:16:12.551083 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3a84e2da-773c-4bf7-8f99-a84276791e78-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-9hpsr\" (UID: \"3a84e2da-773c-4bf7-8f99-a84276791e78\") " pod="openstack/ssh-known-hosts-edpm-deployment-9hpsr" Nov 29 05:16:12 crc kubenswrapper[4799]: I1129 05:16:12.654105 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xx2st\" (UniqueName: \"kubernetes.io/projected/3a84e2da-773c-4bf7-8f99-a84276791e78-kube-api-access-xx2st\") pod \"ssh-known-hosts-edpm-deployment-9hpsr\" (UID: \"3a84e2da-773c-4bf7-8f99-a84276791e78\") " pod="openstack/ssh-known-hosts-edpm-deployment-9hpsr" Nov 29 05:16:12 crc kubenswrapper[4799]: I1129 05:16:12.654643 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3a84e2da-773c-4bf7-8f99-a84276791e78-ceph\") pod \"ssh-known-hosts-edpm-deployment-9hpsr\" (UID: \"3a84e2da-773c-4bf7-8f99-a84276791e78\") " pod="openstack/ssh-known-hosts-edpm-deployment-9hpsr" Nov 29 05:16:12 crc kubenswrapper[4799]: I1129 05:16:12.654726 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/3a84e2da-773c-4bf7-8f99-a84276791e78-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-9hpsr\" (UID: \"3a84e2da-773c-4bf7-8f99-a84276791e78\") " pod="openstack/ssh-known-hosts-edpm-deployment-9hpsr" Nov 29 05:16:12 crc kubenswrapper[4799]: I1129 05:16:12.654953 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3a84e2da-773c-4bf7-8f99-a84276791e78-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-9hpsr\" (UID: \"3a84e2da-773c-4bf7-8f99-a84276791e78\") " pod="openstack/ssh-known-hosts-edpm-deployment-9hpsr" Nov 29 05:16:12 crc kubenswrapper[4799]: I1129 05:16:12.661824 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/3a84e2da-773c-4bf7-8f99-a84276791e78-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-9hpsr\" (UID: \"3a84e2da-773c-4bf7-8f99-a84276791e78\") " pod="openstack/ssh-known-hosts-edpm-deployment-9hpsr" Nov 29 05:16:12 crc kubenswrapper[4799]: I1129 05:16:12.662767 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3a84e2da-773c-4bf7-8f99-a84276791e78-ceph\") pod \"ssh-known-hosts-edpm-deployment-9hpsr\" (UID: \"3a84e2da-773c-4bf7-8f99-a84276791e78\") " pod="openstack/ssh-known-hosts-edpm-deployment-9hpsr" Nov 29 05:16:12 crc kubenswrapper[4799]: I1129 05:16:12.665670 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3a84e2da-773c-4bf7-8f99-a84276791e78-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-9hpsr\" (UID: \"3a84e2da-773c-4bf7-8f99-a84276791e78\") " pod="openstack/ssh-known-hosts-edpm-deployment-9hpsr" Nov 29 05:16:12 crc kubenswrapper[4799]: I1129 05:16:12.686707 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xx2st\" (UniqueName: \"kubernetes.io/projected/3a84e2da-773c-4bf7-8f99-a84276791e78-kube-api-access-xx2st\") pod \"ssh-known-hosts-edpm-deployment-9hpsr\" (UID: \"3a84e2da-773c-4bf7-8f99-a84276791e78\") " pod="openstack/ssh-known-hosts-edpm-deployment-9hpsr" Nov 29 05:16:12 crc kubenswrapper[4799]: I1129 05:16:12.759145 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-9hpsr" Nov 29 05:16:13 crc kubenswrapper[4799]: I1129 05:16:13.194011 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-9hpsr"] Nov 29 05:16:13 crc kubenswrapper[4799]: I1129 05:16:13.278977 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-9hpsr" event={"ID":"3a84e2da-773c-4bf7-8f99-a84276791e78","Type":"ContainerStarted","Data":"49f3d934d810ed7b55cb932a3a733917b6abdc32932641c74a7e4cb7f1af88d1"} Nov 29 05:16:14 crc kubenswrapper[4799]: I1129 05:16:14.293973 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-9hpsr" event={"ID":"3a84e2da-773c-4bf7-8f99-a84276791e78","Type":"ContainerStarted","Data":"8777dffa6d52ffb439f0576c99469fa8ee5f2e2f73ac402f0a0b8d731c55bc72"} Nov 29 05:16:14 crc kubenswrapper[4799]: I1129 05:16:14.325396 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-9hpsr" podStartSLOduration=1.5798694009999998 podStartE2EDuration="2.325367393s" podCreationTimestamp="2025-11-29 05:16:12 +0000 UTC" firstStartedPulling="2025-11-29 05:16:13.204747502 +0000 UTC m=+2248.847677912" lastFinishedPulling="2025-11-29 05:16:13.950245464 +0000 UTC m=+2249.593175904" observedRunningTime="2025-11-29 05:16:14.317745715 +0000 UTC m=+2249.960676165" watchObservedRunningTime="2025-11-29 05:16:14.325367393 +0000 UTC m=+2249.968297833" Nov 29 05:16:20 crc kubenswrapper[4799]: I1129 05:16:20.659716 4799 scope.go:117] "RemoveContainer" containerID="d0281928bf1426881365580d5726280995a377a80c626b0654de38a8828878e1" Nov 29 05:16:20 crc kubenswrapper[4799]: E1129 05:16:20.661325 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:16:25 crc kubenswrapper[4799]: I1129 05:16:25.444743 4799 generic.go:334] "Generic (PLEG): container finished" podID="3a84e2da-773c-4bf7-8f99-a84276791e78" containerID="8777dffa6d52ffb439f0576c99469fa8ee5f2e2f73ac402f0a0b8d731c55bc72" exitCode=0 Nov 29 05:16:25 crc kubenswrapper[4799]: I1129 05:16:25.444834 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-9hpsr" event={"ID":"3a84e2da-773c-4bf7-8f99-a84276791e78","Type":"ContainerDied","Data":"8777dffa6d52ffb439f0576c99469fa8ee5f2e2f73ac402f0a0b8d731c55bc72"} Nov 29 05:16:26 crc kubenswrapper[4799]: I1129 05:16:26.959149 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-9hpsr" Nov 29 05:16:27 crc kubenswrapper[4799]: I1129 05:16:27.029155 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3a84e2da-773c-4bf7-8f99-a84276791e78-ssh-key-openstack-edpm-ipam\") pod \"3a84e2da-773c-4bf7-8f99-a84276791e78\" (UID: \"3a84e2da-773c-4bf7-8f99-a84276791e78\") " Nov 29 05:16:27 crc kubenswrapper[4799]: I1129 05:16:27.029677 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/3a84e2da-773c-4bf7-8f99-a84276791e78-inventory-0\") pod \"3a84e2da-773c-4bf7-8f99-a84276791e78\" (UID: \"3a84e2da-773c-4bf7-8f99-a84276791e78\") " Nov 29 05:16:27 crc kubenswrapper[4799]: I1129 05:16:27.029749 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3a84e2da-773c-4bf7-8f99-a84276791e78-ceph\") pod \"3a84e2da-773c-4bf7-8f99-a84276791e78\" (UID: \"3a84e2da-773c-4bf7-8f99-a84276791e78\") " Nov 29 05:16:27 crc kubenswrapper[4799]: I1129 05:16:27.029806 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xx2st\" (UniqueName: \"kubernetes.io/projected/3a84e2da-773c-4bf7-8f99-a84276791e78-kube-api-access-xx2st\") pod \"3a84e2da-773c-4bf7-8f99-a84276791e78\" (UID: \"3a84e2da-773c-4bf7-8f99-a84276791e78\") " Nov 29 05:16:27 crc kubenswrapper[4799]: I1129 05:16:27.040190 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a84e2da-773c-4bf7-8f99-a84276791e78-ceph" (OuterVolumeSpecName: "ceph") pod "3a84e2da-773c-4bf7-8f99-a84276791e78" (UID: "3a84e2da-773c-4bf7-8f99-a84276791e78"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:16:27 crc kubenswrapper[4799]: I1129 05:16:27.040271 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a84e2da-773c-4bf7-8f99-a84276791e78-kube-api-access-xx2st" (OuterVolumeSpecName: "kube-api-access-xx2st") pod "3a84e2da-773c-4bf7-8f99-a84276791e78" (UID: "3a84e2da-773c-4bf7-8f99-a84276791e78"). InnerVolumeSpecName "kube-api-access-xx2st". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:16:27 crc kubenswrapper[4799]: I1129 05:16:27.074656 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a84e2da-773c-4bf7-8f99-a84276791e78-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "3a84e2da-773c-4bf7-8f99-a84276791e78" (UID: "3a84e2da-773c-4bf7-8f99-a84276791e78"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:16:27 crc kubenswrapper[4799]: I1129 05:16:27.077151 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a84e2da-773c-4bf7-8f99-a84276791e78-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "3a84e2da-773c-4bf7-8f99-a84276791e78" (UID: "3a84e2da-773c-4bf7-8f99-a84276791e78"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:16:27 crc kubenswrapper[4799]: I1129 05:16:27.133063 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3a84e2da-773c-4bf7-8f99-a84276791e78-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 29 05:16:27 crc kubenswrapper[4799]: I1129 05:16:27.133111 4799 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/3a84e2da-773c-4bf7-8f99-a84276791e78-inventory-0\") on node \"crc\" DevicePath \"\"" Nov 29 05:16:27 crc kubenswrapper[4799]: I1129 05:16:27.133125 4799 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3a84e2da-773c-4bf7-8f99-a84276791e78-ceph\") on node \"crc\" DevicePath \"\"" Nov 29 05:16:27 crc kubenswrapper[4799]: I1129 05:16:27.133138 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xx2st\" (UniqueName: \"kubernetes.io/projected/3a84e2da-773c-4bf7-8f99-a84276791e78-kube-api-access-xx2st\") on node \"crc\" DevicePath \"\"" Nov 29 05:16:27 crc kubenswrapper[4799]: I1129 05:16:27.466395 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-9hpsr" event={"ID":"3a84e2da-773c-4bf7-8f99-a84276791e78","Type":"ContainerDied","Data":"49f3d934d810ed7b55cb932a3a733917b6abdc32932641c74a7e4cb7f1af88d1"} Nov 29 05:16:27 crc kubenswrapper[4799]: I1129 05:16:27.466457 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="49f3d934d810ed7b55cb932a3a733917b6abdc32932641c74a7e4cb7f1af88d1" Nov 29 05:16:27 crc kubenswrapper[4799]: I1129 05:16:27.466498 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-9hpsr" Nov 29 05:16:27 crc kubenswrapper[4799]: I1129 05:16:27.579699 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-h4p62"] Nov 29 05:16:27 crc kubenswrapper[4799]: E1129 05:16:27.580549 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a84e2da-773c-4bf7-8f99-a84276791e78" containerName="ssh-known-hosts-edpm-deployment" Nov 29 05:16:27 crc kubenswrapper[4799]: I1129 05:16:27.580593 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a84e2da-773c-4bf7-8f99-a84276791e78" containerName="ssh-known-hosts-edpm-deployment" Nov 29 05:16:27 crc kubenswrapper[4799]: I1129 05:16:27.581145 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a84e2da-773c-4bf7-8f99-a84276791e78" containerName="ssh-known-hosts-edpm-deployment" Nov 29 05:16:27 crc kubenswrapper[4799]: I1129 05:16:27.582516 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-h4p62" Nov 29 05:16:27 crc kubenswrapper[4799]: I1129 05:16:27.586221 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 29 05:16:27 crc kubenswrapper[4799]: I1129 05:16:27.586266 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 29 05:16:27 crc kubenswrapper[4799]: I1129 05:16:27.586395 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 29 05:16:27 crc kubenswrapper[4799]: I1129 05:16:27.586720 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 29 05:16:27 crc kubenswrapper[4799]: I1129 05:16:27.587250 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ljwtk" Nov 29 05:16:27 crc kubenswrapper[4799]: I1129 05:16:27.598819 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-h4p62"] Nov 29 05:16:27 crc kubenswrapper[4799]: I1129 05:16:27.758261 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/828374a6-f506-4e41-83fe-32bf0f6a4ed3-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-h4p62\" (UID: \"828374a6-f506-4e41-83fe-32bf0f6a4ed3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-h4p62" Nov 29 05:16:27 crc kubenswrapper[4799]: I1129 05:16:27.758338 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4whd\" (UniqueName: \"kubernetes.io/projected/828374a6-f506-4e41-83fe-32bf0f6a4ed3-kube-api-access-m4whd\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-h4p62\" (UID: \"828374a6-f506-4e41-83fe-32bf0f6a4ed3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-h4p62" Nov 29 05:16:27 crc kubenswrapper[4799]: I1129 05:16:27.758369 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/828374a6-f506-4e41-83fe-32bf0f6a4ed3-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-h4p62\" (UID: \"828374a6-f506-4e41-83fe-32bf0f6a4ed3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-h4p62" Nov 29 05:16:27 crc kubenswrapper[4799]: I1129 05:16:27.758405 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/828374a6-f506-4e41-83fe-32bf0f6a4ed3-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-h4p62\" (UID: \"828374a6-f506-4e41-83fe-32bf0f6a4ed3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-h4p62" Nov 29 05:16:27 crc kubenswrapper[4799]: I1129 05:16:27.860599 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/828374a6-f506-4e41-83fe-32bf0f6a4ed3-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-h4p62\" (UID: \"828374a6-f506-4e41-83fe-32bf0f6a4ed3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-h4p62" Nov 29 05:16:27 crc kubenswrapper[4799]: I1129 05:16:27.860682 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4whd\" (UniqueName: \"kubernetes.io/projected/828374a6-f506-4e41-83fe-32bf0f6a4ed3-kube-api-access-m4whd\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-h4p62\" (UID: \"828374a6-f506-4e41-83fe-32bf0f6a4ed3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-h4p62" Nov 29 05:16:27 crc kubenswrapper[4799]: I1129 05:16:27.860709 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/828374a6-f506-4e41-83fe-32bf0f6a4ed3-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-h4p62\" (UID: \"828374a6-f506-4e41-83fe-32bf0f6a4ed3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-h4p62" Nov 29 05:16:27 crc kubenswrapper[4799]: I1129 05:16:27.860764 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/828374a6-f506-4e41-83fe-32bf0f6a4ed3-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-h4p62\" (UID: \"828374a6-f506-4e41-83fe-32bf0f6a4ed3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-h4p62" Nov 29 05:16:27 crc kubenswrapper[4799]: I1129 05:16:27.867051 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/828374a6-f506-4e41-83fe-32bf0f6a4ed3-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-h4p62\" (UID: \"828374a6-f506-4e41-83fe-32bf0f6a4ed3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-h4p62" Nov 29 05:16:27 crc kubenswrapper[4799]: I1129 05:16:27.876841 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/828374a6-f506-4e41-83fe-32bf0f6a4ed3-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-h4p62\" (UID: \"828374a6-f506-4e41-83fe-32bf0f6a4ed3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-h4p62" Nov 29 05:16:27 crc kubenswrapper[4799]: I1129 05:16:27.877028 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/828374a6-f506-4e41-83fe-32bf0f6a4ed3-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-h4p62\" (UID: \"828374a6-f506-4e41-83fe-32bf0f6a4ed3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-h4p62" Nov 29 05:16:27 crc kubenswrapper[4799]: I1129 05:16:27.883159 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4whd\" (UniqueName: \"kubernetes.io/projected/828374a6-f506-4e41-83fe-32bf0f6a4ed3-kube-api-access-m4whd\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-h4p62\" (UID: \"828374a6-f506-4e41-83fe-32bf0f6a4ed3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-h4p62" Nov 29 05:16:27 crc kubenswrapper[4799]: I1129 05:16:27.909881 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-h4p62" Nov 29 05:16:28 crc kubenswrapper[4799]: I1129 05:16:28.555229 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-h4p62"] Nov 29 05:16:29 crc kubenswrapper[4799]: I1129 05:16:29.492106 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-h4p62" event={"ID":"828374a6-f506-4e41-83fe-32bf0f6a4ed3","Type":"ContainerStarted","Data":"015473a7d0b481619dea7f763d6ea06e660529020eb55e567193ec880dd41c2d"} Nov 29 05:16:29 crc kubenswrapper[4799]: I1129 05:16:29.492768 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-h4p62" event={"ID":"828374a6-f506-4e41-83fe-32bf0f6a4ed3","Type":"ContainerStarted","Data":"55d48f925624f0365719636c6d37553184c621568d5068a4689f106afd9f5f8f"} Nov 29 05:16:33 crc kubenswrapper[4799]: I1129 05:16:33.660132 4799 scope.go:117] "RemoveContainer" containerID="d0281928bf1426881365580d5726280995a377a80c626b0654de38a8828878e1" Nov 29 05:16:33 crc kubenswrapper[4799]: E1129 05:16:33.661682 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:16:38 crc kubenswrapper[4799]: I1129 05:16:38.597943 4799 generic.go:334] "Generic (PLEG): container finished" podID="828374a6-f506-4e41-83fe-32bf0f6a4ed3" containerID="015473a7d0b481619dea7f763d6ea06e660529020eb55e567193ec880dd41c2d" exitCode=0 Nov 29 05:16:38 crc kubenswrapper[4799]: I1129 05:16:38.598053 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-h4p62" event={"ID":"828374a6-f506-4e41-83fe-32bf0f6a4ed3","Type":"ContainerDied","Data":"015473a7d0b481619dea7f763d6ea06e660529020eb55e567193ec880dd41c2d"} Nov 29 05:16:40 crc kubenswrapper[4799]: I1129 05:16:40.125431 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-h4p62" Nov 29 05:16:40 crc kubenswrapper[4799]: I1129 05:16:40.288653 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m4whd\" (UniqueName: \"kubernetes.io/projected/828374a6-f506-4e41-83fe-32bf0f6a4ed3-kube-api-access-m4whd\") pod \"828374a6-f506-4e41-83fe-32bf0f6a4ed3\" (UID: \"828374a6-f506-4e41-83fe-32bf0f6a4ed3\") " Nov 29 05:16:40 crc kubenswrapper[4799]: I1129 05:16:40.289078 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/828374a6-f506-4e41-83fe-32bf0f6a4ed3-ssh-key\") pod \"828374a6-f506-4e41-83fe-32bf0f6a4ed3\" (UID: \"828374a6-f506-4e41-83fe-32bf0f6a4ed3\") " Nov 29 05:16:40 crc kubenswrapper[4799]: I1129 05:16:40.289231 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/828374a6-f506-4e41-83fe-32bf0f6a4ed3-ceph\") pod \"828374a6-f506-4e41-83fe-32bf0f6a4ed3\" (UID: \"828374a6-f506-4e41-83fe-32bf0f6a4ed3\") " Nov 29 05:16:40 crc kubenswrapper[4799]: I1129 05:16:40.289348 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/828374a6-f506-4e41-83fe-32bf0f6a4ed3-inventory\") pod \"828374a6-f506-4e41-83fe-32bf0f6a4ed3\" (UID: \"828374a6-f506-4e41-83fe-32bf0f6a4ed3\") " Nov 29 05:16:40 crc kubenswrapper[4799]: I1129 05:16:40.296969 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/828374a6-f506-4e41-83fe-32bf0f6a4ed3-ceph" (OuterVolumeSpecName: "ceph") pod "828374a6-f506-4e41-83fe-32bf0f6a4ed3" (UID: "828374a6-f506-4e41-83fe-32bf0f6a4ed3"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:16:40 crc kubenswrapper[4799]: I1129 05:16:40.297630 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/828374a6-f506-4e41-83fe-32bf0f6a4ed3-kube-api-access-m4whd" (OuterVolumeSpecName: "kube-api-access-m4whd") pod "828374a6-f506-4e41-83fe-32bf0f6a4ed3" (UID: "828374a6-f506-4e41-83fe-32bf0f6a4ed3"). InnerVolumeSpecName "kube-api-access-m4whd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:16:40 crc kubenswrapper[4799]: I1129 05:16:40.330514 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/828374a6-f506-4e41-83fe-32bf0f6a4ed3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "828374a6-f506-4e41-83fe-32bf0f6a4ed3" (UID: "828374a6-f506-4e41-83fe-32bf0f6a4ed3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:16:40 crc kubenswrapper[4799]: I1129 05:16:40.351569 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/828374a6-f506-4e41-83fe-32bf0f6a4ed3-inventory" (OuterVolumeSpecName: "inventory") pod "828374a6-f506-4e41-83fe-32bf0f6a4ed3" (UID: "828374a6-f506-4e41-83fe-32bf0f6a4ed3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:16:40 crc kubenswrapper[4799]: I1129 05:16:40.392277 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m4whd\" (UniqueName: \"kubernetes.io/projected/828374a6-f506-4e41-83fe-32bf0f6a4ed3-kube-api-access-m4whd\") on node \"crc\" DevicePath \"\"" Nov 29 05:16:40 crc kubenswrapper[4799]: I1129 05:16:40.392329 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/828374a6-f506-4e41-83fe-32bf0f6a4ed3-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 29 05:16:40 crc kubenswrapper[4799]: I1129 05:16:40.392349 4799 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/828374a6-f506-4e41-83fe-32bf0f6a4ed3-ceph\") on node \"crc\" DevicePath \"\"" Nov 29 05:16:40 crc kubenswrapper[4799]: I1129 05:16:40.392367 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/828374a6-f506-4e41-83fe-32bf0f6a4ed3-inventory\") on node \"crc\" DevicePath \"\"" Nov 29 05:16:40 crc kubenswrapper[4799]: I1129 05:16:40.624377 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-h4p62" event={"ID":"828374a6-f506-4e41-83fe-32bf0f6a4ed3","Type":"ContainerDied","Data":"55d48f925624f0365719636c6d37553184c621568d5068a4689f106afd9f5f8f"} Nov 29 05:16:40 crc kubenswrapper[4799]: I1129 05:16:40.624947 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="55d48f925624f0365719636c6d37553184c621568d5068a4689f106afd9f5f8f" Nov 29 05:16:40 crc kubenswrapper[4799]: I1129 05:16:40.624476 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-h4p62" Nov 29 05:16:40 crc kubenswrapper[4799]: I1129 05:16:40.754732 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-d68wr"] Nov 29 05:16:40 crc kubenswrapper[4799]: E1129 05:16:40.755622 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="828374a6-f506-4e41-83fe-32bf0f6a4ed3" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 29 05:16:40 crc kubenswrapper[4799]: I1129 05:16:40.755646 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="828374a6-f506-4e41-83fe-32bf0f6a4ed3" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 29 05:16:40 crc kubenswrapper[4799]: I1129 05:16:40.755851 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="828374a6-f506-4e41-83fe-32bf0f6a4ed3" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 29 05:16:40 crc kubenswrapper[4799]: I1129 05:16:40.757687 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-d68wr" Nov 29 05:16:40 crc kubenswrapper[4799]: I1129 05:16:40.760268 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 29 05:16:40 crc kubenswrapper[4799]: I1129 05:16:40.761500 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 29 05:16:40 crc kubenswrapper[4799]: I1129 05:16:40.761741 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ljwtk" Nov 29 05:16:40 crc kubenswrapper[4799]: I1129 05:16:40.761951 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 29 05:16:40 crc kubenswrapper[4799]: I1129 05:16:40.763827 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 29 05:16:40 crc kubenswrapper[4799]: I1129 05:16:40.770065 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-d68wr"] Nov 29 05:16:40 crc kubenswrapper[4799]: I1129 05:16:40.906222 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0a726ce8-7d0d-4906-a2a4-e6f146dc4635-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-d68wr\" (UID: \"0a726ce8-7d0d-4906-a2a4-e6f146dc4635\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-d68wr" Nov 29 05:16:40 crc kubenswrapper[4799]: I1129 05:16:40.906317 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgd94\" (UniqueName: \"kubernetes.io/projected/0a726ce8-7d0d-4906-a2a4-e6f146dc4635-kube-api-access-zgd94\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-d68wr\" (UID: \"0a726ce8-7d0d-4906-a2a4-e6f146dc4635\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-d68wr" Nov 29 05:16:40 crc kubenswrapper[4799]: I1129 05:16:40.906357 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0a726ce8-7d0d-4906-a2a4-e6f146dc4635-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-d68wr\" (UID: \"0a726ce8-7d0d-4906-a2a4-e6f146dc4635\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-d68wr" Nov 29 05:16:40 crc kubenswrapper[4799]: I1129 05:16:40.906472 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0a726ce8-7d0d-4906-a2a4-e6f146dc4635-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-d68wr\" (UID: \"0a726ce8-7d0d-4906-a2a4-e6f146dc4635\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-d68wr" Nov 29 05:16:41 crc kubenswrapper[4799]: I1129 05:16:41.007865 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0a726ce8-7d0d-4906-a2a4-e6f146dc4635-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-d68wr\" (UID: \"0a726ce8-7d0d-4906-a2a4-e6f146dc4635\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-d68wr" Nov 29 05:16:41 crc kubenswrapper[4799]: I1129 05:16:41.007944 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0a726ce8-7d0d-4906-a2a4-e6f146dc4635-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-d68wr\" (UID: \"0a726ce8-7d0d-4906-a2a4-e6f146dc4635\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-d68wr" Nov 29 05:16:41 crc kubenswrapper[4799]: I1129 05:16:41.008003 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgd94\" (UniqueName: \"kubernetes.io/projected/0a726ce8-7d0d-4906-a2a4-e6f146dc4635-kube-api-access-zgd94\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-d68wr\" (UID: \"0a726ce8-7d0d-4906-a2a4-e6f146dc4635\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-d68wr" Nov 29 05:16:41 crc kubenswrapper[4799]: I1129 05:16:41.008039 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0a726ce8-7d0d-4906-a2a4-e6f146dc4635-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-d68wr\" (UID: \"0a726ce8-7d0d-4906-a2a4-e6f146dc4635\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-d68wr" Nov 29 05:16:41 crc kubenswrapper[4799]: I1129 05:16:41.015714 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0a726ce8-7d0d-4906-a2a4-e6f146dc4635-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-d68wr\" (UID: \"0a726ce8-7d0d-4906-a2a4-e6f146dc4635\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-d68wr" Nov 29 05:16:41 crc kubenswrapper[4799]: I1129 05:16:41.018249 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0a726ce8-7d0d-4906-a2a4-e6f146dc4635-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-d68wr\" (UID: \"0a726ce8-7d0d-4906-a2a4-e6f146dc4635\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-d68wr" Nov 29 05:16:41 crc kubenswrapper[4799]: I1129 05:16:41.018376 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0a726ce8-7d0d-4906-a2a4-e6f146dc4635-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-d68wr\" (UID: \"0a726ce8-7d0d-4906-a2a4-e6f146dc4635\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-d68wr" Nov 29 05:16:41 crc kubenswrapper[4799]: I1129 05:16:41.030062 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgd94\" (UniqueName: \"kubernetes.io/projected/0a726ce8-7d0d-4906-a2a4-e6f146dc4635-kube-api-access-zgd94\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-d68wr\" (UID: \"0a726ce8-7d0d-4906-a2a4-e6f146dc4635\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-d68wr" Nov 29 05:16:41 crc kubenswrapper[4799]: I1129 05:16:41.129339 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-d68wr" Nov 29 05:16:41 crc kubenswrapper[4799]: I1129 05:16:41.515869 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-d68wr"] Nov 29 05:16:41 crc kubenswrapper[4799]: I1129 05:16:41.637270 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-d68wr" event={"ID":"0a726ce8-7d0d-4906-a2a4-e6f146dc4635","Type":"ContainerStarted","Data":"3482538e5027b427a3b1b4921aabfa2050c13d7074721c599bda8ed0df28f205"} Nov 29 05:16:42 crc kubenswrapper[4799]: I1129 05:16:42.703517 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-d68wr" podStartSLOduration=2.213982786 podStartE2EDuration="2.703484346s" podCreationTimestamp="2025-11-29 05:16:40 +0000 UTC" firstStartedPulling="2025-11-29 05:16:41.540332735 +0000 UTC m=+2277.183263155" lastFinishedPulling="2025-11-29 05:16:42.029834315 +0000 UTC m=+2277.672764715" observedRunningTime="2025-11-29 05:16:42.695747186 +0000 UTC m=+2278.338677666" watchObservedRunningTime="2025-11-29 05:16:42.703484346 +0000 UTC m=+2278.346414756" Nov 29 05:16:42 crc kubenswrapper[4799]: I1129 05:16:42.711464 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-d68wr" event={"ID":"0a726ce8-7d0d-4906-a2a4-e6f146dc4635","Type":"ContainerStarted","Data":"00f0fa3a95c602db5d3db406f224957795975438c30ce8639e1acf18f855bdda"} Nov 29 05:16:44 crc kubenswrapper[4799]: I1129 05:16:44.666880 4799 scope.go:117] "RemoveContainer" containerID="d0281928bf1426881365580d5726280995a377a80c626b0654de38a8828878e1" Nov 29 05:16:44 crc kubenswrapper[4799]: E1129 05:16:44.667478 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:16:52 crc kubenswrapper[4799]: I1129 05:16:52.793324 4799 generic.go:334] "Generic (PLEG): container finished" podID="0a726ce8-7d0d-4906-a2a4-e6f146dc4635" containerID="00f0fa3a95c602db5d3db406f224957795975438c30ce8639e1acf18f855bdda" exitCode=0 Nov 29 05:16:52 crc kubenswrapper[4799]: I1129 05:16:52.793452 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-d68wr" event={"ID":"0a726ce8-7d0d-4906-a2a4-e6f146dc4635","Type":"ContainerDied","Data":"00f0fa3a95c602db5d3db406f224957795975438c30ce8639e1acf18f855bdda"} Nov 29 05:16:54 crc kubenswrapper[4799]: I1129 05:16:54.281642 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-d68wr" Nov 29 05:16:54 crc kubenswrapper[4799]: I1129 05:16:54.351717 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0a726ce8-7d0d-4906-a2a4-e6f146dc4635-inventory\") pod \"0a726ce8-7d0d-4906-a2a4-e6f146dc4635\" (UID: \"0a726ce8-7d0d-4906-a2a4-e6f146dc4635\") " Nov 29 05:16:54 crc kubenswrapper[4799]: I1129 05:16:54.351875 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0a726ce8-7d0d-4906-a2a4-e6f146dc4635-ceph\") pod \"0a726ce8-7d0d-4906-a2a4-e6f146dc4635\" (UID: \"0a726ce8-7d0d-4906-a2a4-e6f146dc4635\") " Nov 29 05:16:54 crc kubenswrapper[4799]: I1129 05:16:54.352065 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgd94\" (UniqueName: \"kubernetes.io/projected/0a726ce8-7d0d-4906-a2a4-e6f146dc4635-kube-api-access-zgd94\") pod \"0a726ce8-7d0d-4906-a2a4-e6f146dc4635\" (UID: \"0a726ce8-7d0d-4906-a2a4-e6f146dc4635\") " Nov 29 05:16:54 crc kubenswrapper[4799]: I1129 05:16:54.352124 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0a726ce8-7d0d-4906-a2a4-e6f146dc4635-ssh-key\") pod \"0a726ce8-7d0d-4906-a2a4-e6f146dc4635\" (UID: \"0a726ce8-7d0d-4906-a2a4-e6f146dc4635\") " Nov 29 05:16:54 crc kubenswrapper[4799]: I1129 05:16:54.360465 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a726ce8-7d0d-4906-a2a4-e6f146dc4635-ceph" (OuterVolumeSpecName: "ceph") pod "0a726ce8-7d0d-4906-a2a4-e6f146dc4635" (UID: "0a726ce8-7d0d-4906-a2a4-e6f146dc4635"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:16:54 crc kubenswrapper[4799]: I1129 05:16:54.362384 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a726ce8-7d0d-4906-a2a4-e6f146dc4635-kube-api-access-zgd94" (OuterVolumeSpecName: "kube-api-access-zgd94") pod "0a726ce8-7d0d-4906-a2a4-e6f146dc4635" (UID: "0a726ce8-7d0d-4906-a2a4-e6f146dc4635"). InnerVolumeSpecName "kube-api-access-zgd94". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:16:54 crc kubenswrapper[4799]: I1129 05:16:54.384873 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a726ce8-7d0d-4906-a2a4-e6f146dc4635-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0a726ce8-7d0d-4906-a2a4-e6f146dc4635" (UID: "0a726ce8-7d0d-4906-a2a4-e6f146dc4635"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:16:54 crc kubenswrapper[4799]: I1129 05:16:54.407717 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a726ce8-7d0d-4906-a2a4-e6f146dc4635-inventory" (OuterVolumeSpecName: "inventory") pod "0a726ce8-7d0d-4906-a2a4-e6f146dc4635" (UID: "0a726ce8-7d0d-4906-a2a4-e6f146dc4635"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:16:54 crc kubenswrapper[4799]: I1129 05:16:54.456103 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0a726ce8-7d0d-4906-a2a4-e6f146dc4635-inventory\") on node \"crc\" DevicePath \"\"" Nov 29 05:16:54 crc kubenswrapper[4799]: I1129 05:16:54.456169 4799 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0a726ce8-7d0d-4906-a2a4-e6f146dc4635-ceph\") on node \"crc\" DevicePath \"\"" Nov 29 05:16:54 crc kubenswrapper[4799]: I1129 05:16:54.456192 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgd94\" (UniqueName: \"kubernetes.io/projected/0a726ce8-7d0d-4906-a2a4-e6f146dc4635-kube-api-access-zgd94\") on node \"crc\" DevicePath \"\"" Nov 29 05:16:54 crc kubenswrapper[4799]: I1129 05:16:54.456212 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0a726ce8-7d0d-4906-a2a4-e6f146dc4635-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 29 05:16:54 crc kubenswrapper[4799]: I1129 05:16:54.838281 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-d68wr" event={"ID":"0a726ce8-7d0d-4906-a2a4-e6f146dc4635","Type":"ContainerDied","Data":"3482538e5027b427a3b1b4921aabfa2050c13d7074721c599bda8ed0df28f205"} Nov 29 05:16:54 crc kubenswrapper[4799]: I1129 05:16:54.838366 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3482538e5027b427a3b1b4921aabfa2050c13d7074721c599bda8ed0df28f205" Nov 29 05:16:54 crc kubenswrapper[4799]: I1129 05:16:54.838485 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-d68wr" Nov 29 05:16:54 crc kubenswrapper[4799]: I1129 05:16:54.923291 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv"] Nov 29 05:16:54 crc kubenswrapper[4799]: E1129 05:16:54.924412 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a726ce8-7d0d-4906-a2a4-e6f146dc4635" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 29 05:16:54 crc kubenswrapper[4799]: I1129 05:16:54.924438 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a726ce8-7d0d-4906-a2a4-e6f146dc4635" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 29 05:16:54 crc kubenswrapper[4799]: I1129 05:16:54.924708 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a726ce8-7d0d-4906-a2a4-e6f146dc4635" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 29 05:16:54 crc kubenswrapper[4799]: I1129 05:16:54.925683 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv" Nov 29 05:16:54 crc kubenswrapper[4799]: I1129 05:16:54.929989 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Nov 29 05:16:54 crc kubenswrapper[4799]: I1129 05:16:54.930073 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Nov 29 05:16:54 crc kubenswrapper[4799]: I1129 05:16:54.930220 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Nov 29 05:16:54 crc kubenswrapper[4799]: I1129 05:16:54.930310 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 29 05:16:54 crc kubenswrapper[4799]: I1129 05:16:54.930350 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 29 05:16:54 crc kubenswrapper[4799]: I1129 05:16:54.930601 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ljwtk" Nov 29 05:16:54 crc kubenswrapper[4799]: I1129 05:16:54.930639 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 29 05:16:54 crc kubenswrapper[4799]: I1129 05:16:54.930690 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 29 05:16:54 crc kubenswrapper[4799]: I1129 05:16:54.943864 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv"] Nov 29 05:16:55 crc kubenswrapper[4799]: I1129 05:16:55.070442 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv" Nov 29 05:16:55 crc kubenswrapper[4799]: I1129 05:16:55.070708 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv" Nov 29 05:16:55 crc kubenswrapper[4799]: I1129 05:16:55.071038 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv" Nov 29 05:16:55 crc kubenswrapper[4799]: I1129 05:16:55.071261 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv" Nov 29 05:16:55 crc kubenswrapper[4799]: I1129 05:16:55.071451 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzjzz\" (UniqueName: \"kubernetes.io/projected/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-kube-api-access-gzjzz\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv" Nov 29 05:16:55 crc kubenswrapper[4799]: I1129 05:16:55.071499 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv" Nov 29 05:16:55 crc kubenswrapper[4799]: I1129 05:16:55.071573 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv" Nov 29 05:16:55 crc kubenswrapper[4799]: I1129 05:16:55.071626 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv" Nov 29 05:16:55 crc kubenswrapper[4799]: I1129 05:16:55.071875 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv" Nov 29 05:16:55 crc kubenswrapper[4799]: I1129 05:16:55.072050 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv" Nov 29 05:16:55 crc kubenswrapper[4799]: I1129 05:16:55.072117 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv" Nov 29 05:16:55 crc kubenswrapper[4799]: I1129 05:16:55.072175 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv" Nov 29 05:16:55 crc kubenswrapper[4799]: I1129 05:16:55.072243 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv" Nov 29 05:16:55 crc kubenswrapper[4799]: I1129 05:16:55.174160 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv" Nov 29 05:16:55 crc kubenswrapper[4799]: I1129 05:16:55.174224 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv" Nov 29 05:16:55 crc kubenswrapper[4799]: I1129 05:16:55.174261 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv" Nov 29 05:16:55 crc kubenswrapper[4799]: I1129 05:16:55.174303 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv" Nov 29 05:16:55 crc kubenswrapper[4799]: I1129 05:16:55.174346 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv" Nov 29 05:16:55 crc kubenswrapper[4799]: I1129 05:16:55.174386 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzjzz\" (UniqueName: \"kubernetes.io/projected/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-kube-api-access-gzjzz\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv" Nov 29 05:16:55 crc kubenswrapper[4799]: I1129 05:16:55.174411 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv" Nov 29 05:16:55 crc kubenswrapper[4799]: I1129 05:16:55.174438 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv" Nov 29 05:16:55 crc kubenswrapper[4799]: I1129 05:16:55.174460 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv" Nov 29 05:16:55 crc kubenswrapper[4799]: I1129 05:16:55.174487 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv" Nov 29 05:16:55 crc kubenswrapper[4799]: I1129 05:16:55.174524 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv" Nov 29 05:16:55 crc kubenswrapper[4799]: I1129 05:16:55.174550 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv" Nov 29 05:16:55 crc kubenswrapper[4799]: I1129 05:16:55.174580 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv" Nov 29 05:16:55 crc kubenswrapper[4799]: I1129 05:16:55.180891 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv" Nov 29 05:16:55 crc kubenswrapper[4799]: I1129 05:16:55.181427 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv" Nov 29 05:16:55 crc kubenswrapper[4799]: I1129 05:16:55.182154 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv" Nov 29 05:16:55 crc kubenswrapper[4799]: I1129 05:16:55.182296 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv" Nov 29 05:16:55 crc kubenswrapper[4799]: I1129 05:16:55.182665 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv" Nov 29 05:16:55 crc kubenswrapper[4799]: I1129 05:16:55.182997 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv" Nov 29 05:16:55 crc kubenswrapper[4799]: I1129 05:16:55.183554 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv" Nov 29 05:16:55 crc kubenswrapper[4799]: I1129 05:16:55.183817 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv" Nov 29 05:16:55 crc kubenswrapper[4799]: I1129 05:16:55.186204 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv" Nov 29 05:16:55 crc kubenswrapper[4799]: I1129 05:16:55.186414 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv" Nov 29 05:16:55 crc kubenswrapper[4799]: I1129 05:16:55.186480 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv" Nov 29 05:16:55 crc kubenswrapper[4799]: I1129 05:16:55.186524 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv" Nov 29 05:16:55 crc kubenswrapper[4799]: I1129 05:16:55.203566 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzjzz\" (UniqueName: \"kubernetes.io/projected/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-kube-api-access-gzjzz\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv" Nov 29 05:16:55 crc kubenswrapper[4799]: I1129 05:16:55.307743 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv" Nov 29 05:16:55 crc kubenswrapper[4799]: I1129 05:16:55.929076 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv"] Nov 29 05:16:55 crc kubenswrapper[4799]: I1129 05:16:55.941951 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 29 05:16:56 crc kubenswrapper[4799]: I1129 05:16:56.886540 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv" event={"ID":"cbad88f8-b661-4b9f-8815-6c0f2df63d7f","Type":"ContainerStarted","Data":"27cef6007bcd3dbc7ebfec6f60d847a1c17d97516eee0922a8a459a742b901e4"} Nov 29 05:16:56 crc kubenswrapper[4799]: I1129 05:16:56.887728 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv" event={"ID":"cbad88f8-b661-4b9f-8815-6c0f2df63d7f","Type":"ContainerStarted","Data":"4f2de3d25947c3e1408f9a176ab5acd79f095f6ab75ddf5c67bc66a99f312d1d"} Nov 29 05:16:56 crc kubenswrapper[4799]: I1129 05:16:56.925710 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv" podStartSLOduration=2.494608641 podStartE2EDuration="2.92567431s" podCreationTimestamp="2025-11-29 05:16:54 +0000 UTC" firstStartedPulling="2025-11-29 05:16:55.93566383 +0000 UTC m=+2291.578594280" lastFinishedPulling="2025-11-29 05:16:56.366729519 +0000 UTC m=+2292.009659949" observedRunningTime="2025-11-29 05:16:56.920560586 +0000 UTC m=+2292.563491026" watchObservedRunningTime="2025-11-29 05:16:56.92567431 +0000 UTC m=+2292.568604750" Nov 29 05:16:57 crc kubenswrapper[4799]: I1129 05:16:57.659660 4799 scope.go:117] "RemoveContainer" containerID="d0281928bf1426881365580d5726280995a377a80c626b0654de38a8828878e1" Nov 29 05:16:57 crc kubenswrapper[4799]: E1129 05:16:57.660121 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:17:08 crc kubenswrapper[4799]: I1129 05:17:08.659257 4799 scope.go:117] "RemoveContainer" containerID="d0281928bf1426881365580d5726280995a377a80c626b0654de38a8828878e1" Nov 29 05:17:08 crc kubenswrapper[4799]: E1129 05:17:08.660867 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:17:23 crc kubenswrapper[4799]: I1129 05:17:23.660455 4799 scope.go:117] "RemoveContainer" containerID="d0281928bf1426881365580d5726280995a377a80c626b0654de38a8828878e1" Nov 29 05:17:23 crc kubenswrapper[4799]: E1129 05:17:23.661929 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:17:33 crc kubenswrapper[4799]: I1129 05:17:33.358624 4799 generic.go:334] "Generic (PLEG): container finished" podID="cbad88f8-b661-4b9f-8815-6c0f2df63d7f" containerID="27cef6007bcd3dbc7ebfec6f60d847a1c17d97516eee0922a8a459a742b901e4" exitCode=0 Nov 29 05:17:33 crc kubenswrapper[4799]: I1129 05:17:33.358763 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv" event={"ID":"cbad88f8-b661-4b9f-8815-6c0f2df63d7f","Type":"ContainerDied","Data":"27cef6007bcd3dbc7ebfec6f60d847a1c17d97516eee0922a8a459a742b901e4"} Nov 29 05:17:34 crc kubenswrapper[4799]: I1129 05:17:34.979375 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv" Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.107537 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gzjzz\" (UniqueName: \"kubernetes.io/projected/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-kube-api-access-gzjzz\") pod \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.107652 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-ceph\") pod \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.107816 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-repo-setup-combined-ca-bundle\") pod \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.107863 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-neutron-metadata-combined-ca-bundle\") pod \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.107912 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-nova-combined-ca-bundle\") pod \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.107937 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-libvirt-combined-ca-bundle\") pod \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.107980 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-bootstrap-combined-ca-bundle\") pod \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.108028 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-ssh-key\") pod \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.109159 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-ovn-combined-ca-bundle\") pod \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.109285 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.109376 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-inventory\") pod \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.109477 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.109568 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-openstack-edpm-ipam-ovn-default-certs-0\") pod \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\" (UID: \"cbad88f8-b661-4b9f-8815-6c0f2df63d7f\") " Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.115920 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "cbad88f8-b661-4b9f-8815-6c0f2df63d7f" (UID: "cbad88f8-b661-4b9f-8815-6c0f2df63d7f"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.117502 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-ceph" (OuterVolumeSpecName: "ceph") pod "cbad88f8-b661-4b9f-8815-6c0f2df63d7f" (UID: "cbad88f8-b661-4b9f-8815-6c0f2df63d7f"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.118813 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "cbad88f8-b661-4b9f-8815-6c0f2df63d7f" (UID: "cbad88f8-b661-4b9f-8815-6c0f2df63d7f"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.119482 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "cbad88f8-b661-4b9f-8815-6c0f2df63d7f" (UID: "cbad88f8-b661-4b9f-8815-6c0f2df63d7f"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.120158 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "cbad88f8-b661-4b9f-8815-6c0f2df63d7f" (UID: "cbad88f8-b661-4b9f-8815-6c0f2df63d7f"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.120260 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-kube-api-access-gzjzz" (OuterVolumeSpecName: "kube-api-access-gzjzz") pod "cbad88f8-b661-4b9f-8815-6c0f2df63d7f" (UID: "cbad88f8-b661-4b9f-8815-6c0f2df63d7f"). InnerVolumeSpecName "kube-api-access-gzjzz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.120379 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "cbad88f8-b661-4b9f-8815-6c0f2df63d7f" (UID: "cbad88f8-b661-4b9f-8815-6c0f2df63d7f"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.120831 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "cbad88f8-b661-4b9f-8815-6c0f2df63d7f" (UID: "cbad88f8-b661-4b9f-8815-6c0f2df63d7f"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.122072 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "cbad88f8-b661-4b9f-8815-6c0f2df63d7f" (UID: "cbad88f8-b661-4b9f-8815-6c0f2df63d7f"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.122631 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "cbad88f8-b661-4b9f-8815-6c0f2df63d7f" (UID: "cbad88f8-b661-4b9f-8815-6c0f2df63d7f"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.123775 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "cbad88f8-b661-4b9f-8815-6c0f2df63d7f" (UID: "cbad88f8-b661-4b9f-8815-6c0f2df63d7f"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.166817 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-inventory" (OuterVolumeSpecName: "inventory") pod "cbad88f8-b661-4b9f-8815-6c0f2df63d7f" (UID: "cbad88f8-b661-4b9f-8815-6c0f2df63d7f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.183685 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "cbad88f8-b661-4b9f-8815-6c0f2df63d7f" (UID: "cbad88f8-b661-4b9f-8815-6c0f2df63d7f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.214538 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gzjzz\" (UniqueName: \"kubernetes.io/projected/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-kube-api-access-gzjzz\") on node \"crc\" DevicePath \"\"" Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.214587 4799 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-ceph\") on node \"crc\" DevicePath \"\"" Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.214603 4799 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.214617 4799 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.214629 4799 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.214641 4799 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.214652 4799 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.214663 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.214673 4799 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.214686 4799 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.214697 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-inventory\") on node \"crc\" DevicePath \"\"" Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.214706 4799 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.214725 4799 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cbad88f8-b661-4b9f-8815-6c0f2df63d7f-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.385268 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv" event={"ID":"cbad88f8-b661-4b9f-8815-6c0f2df63d7f","Type":"ContainerDied","Data":"4f2de3d25947c3e1408f9a176ab5acd79f095f6ab75ddf5c67bc66a99f312d1d"} Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.385342 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4f2de3d25947c3e1408f9a176ab5acd79f095f6ab75ddf5c67bc66a99f312d1d" Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.385445 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv" Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.562256 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-fdll7"] Nov 29 05:17:35 crc kubenswrapper[4799]: E1129 05:17:35.562702 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbad88f8-b661-4b9f-8815-6c0f2df63d7f" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.562722 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbad88f8-b661-4b9f-8815-6c0f2df63d7f" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.562947 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbad88f8-b661-4b9f-8815-6c0f2df63d7f" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.563664 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-fdll7" Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.571923 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.572180 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.572494 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.572194 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.575036 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ljwtk" Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.584705 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-fdll7"] Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.628344 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/41294840-0c22-4c85-ae8f-31f3a00458d4-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-fdll7\" (UID: \"41294840-0c22-4c85-ae8f-31f3a00458d4\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-fdll7" Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.628492 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sp8tw\" (UniqueName: \"kubernetes.io/projected/41294840-0c22-4c85-ae8f-31f3a00458d4-kube-api-access-sp8tw\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-fdll7\" (UID: \"41294840-0c22-4c85-ae8f-31f3a00458d4\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-fdll7" Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.628695 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/41294840-0c22-4c85-ae8f-31f3a00458d4-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-fdll7\" (UID: \"41294840-0c22-4c85-ae8f-31f3a00458d4\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-fdll7" Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.628744 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/41294840-0c22-4c85-ae8f-31f3a00458d4-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-fdll7\" (UID: \"41294840-0c22-4c85-ae8f-31f3a00458d4\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-fdll7" Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.731072 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/41294840-0c22-4c85-ae8f-31f3a00458d4-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-fdll7\" (UID: \"41294840-0c22-4c85-ae8f-31f3a00458d4\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-fdll7" Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.731153 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/41294840-0c22-4c85-ae8f-31f3a00458d4-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-fdll7\" (UID: \"41294840-0c22-4c85-ae8f-31f3a00458d4\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-fdll7" Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.731191 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/41294840-0c22-4c85-ae8f-31f3a00458d4-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-fdll7\" (UID: \"41294840-0c22-4c85-ae8f-31f3a00458d4\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-fdll7" Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.731249 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sp8tw\" (UniqueName: \"kubernetes.io/projected/41294840-0c22-4c85-ae8f-31f3a00458d4-kube-api-access-sp8tw\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-fdll7\" (UID: \"41294840-0c22-4c85-ae8f-31f3a00458d4\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-fdll7" Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.736613 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/41294840-0c22-4c85-ae8f-31f3a00458d4-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-fdll7\" (UID: \"41294840-0c22-4c85-ae8f-31f3a00458d4\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-fdll7" Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.737190 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/41294840-0c22-4c85-ae8f-31f3a00458d4-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-fdll7\" (UID: \"41294840-0c22-4c85-ae8f-31f3a00458d4\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-fdll7" Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.738770 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/41294840-0c22-4c85-ae8f-31f3a00458d4-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-fdll7\" (UID: \"41294840-0c22-4c85-ae8f-31f3a00458d4\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-fdll7" Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.755059 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sp8tw\" (UniqueName: \"kubernetes.io/projected/41294840-0c22-4c85-ae8f-31f3a00458d4-kube-api-access-sp8tw\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-fdll7\" (UID: \"41294840-0c22-4c85-ae8f-31f3a00458d4\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-fdll7" Nov 29 05:17:35 crc kubenswrapper[4799]: I1129 05:17:35.886592 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-fdll7" Nov 29 05:17:36 crc kubenswrapper[4799]: I1129 05:17:36.563277 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-fdll7"] Nov 29 05:17:36 crc kubenswrapper[4799]: W1129 05:17:36.569539 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod41294840_0c22_4c85_ae8f_31f3a00458d4.slice/crio-0f0a112a03dd494ac333be757200d1ff4fb0c9703b3da0416218a5e3f680f4b0 WatchSource:0}: Error finding container 0f0a112a03dd494ac333be757200d1ff4fb0c9703b3da0416218a5e3f680f4b0: Status 404 returned error can't find the container with id 0f0a112a03dd494ac333be757200d1ff4fb0c9703b3da0416218a5e3f680f4b0 Nov 29 05:17:37 crc kubenswrapper[4799]: I1129 05:17:37.415193 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-fdll7" event={"ID":"41294840-0c22-4c85-ae8f-31f3a00458d4","Type":"ContainerStarted","Data":"0f0a112a03dd494ac333be757200d1ff4fb0c9703b3da0416218a5e3f680f4b0"} Nov 29 05:17:37 crc kubenswrapper[4799]: I1129 05:17:37.660301 4799 scope.go:117] "RemoveContainer" containerID="d0281928bf1426881365580d5726280995a377a80c626b0654de38a8828878e1" Nov 29 05:17:37 crc kubenswrapper[4799]: E1129 05:17:37.661764 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:17:38 crc kubenswrapper[4799]: I1129 05:17:38.432486 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-fdll7" event={"ID":"41294840-0c22-4c85-ae8f-31f3a00458d4","Type":"ContainerStarted","Data":"9795708e6fadabdc0828d0d724c5c2d0168f450dea096975a1638549744d5a21"} Nov 29 05:17:38 crc kubenswrapper[4799]: I1129 05:17:38.470662 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-fdll7" podStartSLOduration=2.614054538 podStartE2EDuration="3.47063034s" podCreationTimestamp="2025-11-29 05:17:35 +0000 UTC" firstStartedPulling="2025-11-29 05:17:36.573134131 +0000 UTC m=+2332.216064531" lastFinishedPulling="2025-11-29 05:17:37.429709933 +0000 UTC m=+2333.072640333" observedRunningTime="2025-11-29 05:17:38.463376112 +0000 UTC m=+2334.106306552" watchObservedRunningTime="2025-11-29 05:17:38.47063034 +0000 UTC m=+2334.113560770" Nov 29 05:17:43 crc kubenswrapper[4799]: I1129 05:17:43.493608 4799 generic.go:334] "Generic (PLEG): container finished" podID="41294840-0c22-4c85-ae8f-31f3a00458d4" containerID="9795708e6fadabdc0828d0d724c5c2d0168f450dea096975a1638549744d5a21" exitCode=0 Nov 29 05:17:43 crc kubenswrapper[4799]: I1129 05:17:43.493764 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-fdll7" event={"ID":"41294840-0c22-4c85-ae8f-31f3a00458d4","Type":"ContainerDied","Data":"9795708e6fadabdc0828d0d724c5c2d0168f450dea096975a1638549744d5a21"} Nov 29 05:17:45 crc kubenswrapper[4799]: I1129 05:17:45.019182 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-fdll7" Nov 29 05:17:45 crc kubenswrapper[4799]: I1129 05:17:45.077143 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/41294840-0c22-4c85-ae8f-31f3a00458d4-inventory\") pod \"41294840-0c22-4c85-ae8f-31f3a00458d4\" (UID: \"41294840-0c22-4c85-ae8f-31f3a00458d4\") " Nov 29 05:17:45 crc kubenswrapper[4799]: I1129 05:17:45.077308 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/41294840-0c22-4c85-ae8f-31f3a00458d4-ceph\") pod \"41294840-0c22-4c85-ae8f-31f3a00458d4\" (UID: \"41294840-0c22-4c85-ae8f-31f3a00458d4\") " Nov 29 05:17:45 crc kubenswrapper[4799]: I1129 05:17:45.077350 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sp8tw\" (UniqueName: \"kubernetes.io/projected/41294840-0c22-4c85-ae8f-31f3a00458d4-kube-api-access-sp8tw\") pod \"41294840-0c22-4c85-ae8f-31f3a00458d4\" (UID: \"41294840-0c22-4c85-ae8f-31f3a00458d4\") " Nov 29 05:17:45 crc kubenswrapper[4799]: I1129 05:17:45.077422 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/41294840-0c22-4c85-ae8f-31f3a00458d4-ssh-key\") pod \"41294840-0c22-4c85-ae8f-31f3a00458d4\" (UID: \"41294840-0c22-4c85-ae8f-31f3a00458d4\") " Nov 29 05:17:45 crc kubenswrapper[4799]: I1129 05:17:45.086618 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41294840-0c22-4c85-ae8f-31f3a00458d4-kube-api-access-sp8tw" (OuterVolumeSpecName: "kube-api-access-sp8tw") pod "41294840-0c22-4c85-ae8f-31f3a00458d4" (UID: "41294840-0c22-4c85-ae8f-31f3a00458d4"). InnerVolumeSpecName "kube-api-access-sp8tw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:17:45 crc kubenswrapper[4799]: I1129 05:17:45.091126 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41294840-0c22-4c85-ae8f-31f3a00458d4-ceph" (OuterVolumeSpecName: "ceph") pod "41294840-0c22-4c85-ae8f-31f3a00458d4" (UID: "41294840-0c22-4c85-ae8f-31f3a00458d4"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:17:45 crc kubenswrapper[4799]: I1129 05:17:45.111705 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41294840-0c22-4c85-ae8f-31f3a00458d4-inventory" (OuterVolumeSpecName: "inventory") pod "41294840-0c22-4c85-ae8f-31f3a00458d4" (UID: "41294840-0c22-4c85-ae8f-31f3a00458d4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:17:45 crc kubenswrapper[4799]: I1129 05:17:45.111997 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41294840-0c22-4c85-ae8f-31f3a00458d4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "41294840-0c22-4c85-ae8f-31f3a00458d4" (UID: "41294840-0c22-4c85-ae8f-31f3a00458d4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:17:45 crc kubenswrapper[4799]: I1129 05:17:45.179884 4799 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/41294840-0c22-4c85-ae8f-31f3a00458d4-ceph\") on node \"crc\" DevicePath \"\"" Nov 29 05:17:45 crc kubenswrapper[4799]: I1129 05:17:45.180763 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sp8tw\" (UniqueName: \"kubernetes.io/projected/41294840-0c22-4c85-ae8f-31f3a00458d4-kube-api-access-sp8tw\") on node \"crc\" DevicePath \"\"" Nov 29 05:17:45 crc kubenswrapper[4799]: I1129 05:17:45.180895 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/41294840-0c22-4c85-ae8f-31f3a00458d4-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 29 05:17:45 crc kubenswrapper[4799]: I1129 05:17:45.180972 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/41294840-0c22-4c85-ae8f-31f3a00458d4-inventory\") on node \"crc\" DevicePath \"\"" Nov 29 05:17:45 crc kubenswrapper[4799]: I1129 05:17:45.519314 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-fdll7" event={"ID":"41294840-0c22-4c85-ae8f-31f3a00458d4","Type":"ContainerDied","Data":"0f0a112a03dd494ac333be757200d1ff4fb0c9703b3da0416218a5e3f680f4b0"} Nov 29 05:17:45 crc kubenswrapper[4799]: I1129 05:17:45.519394 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0f0a112a03dd494ac333be757200d1ff4fb0c9703b3da0416218a5e3f680f4b0" Nov 29 05:17:45 crc kubenswrapper[4799]: I1129 05:17:45.519949 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-fdll7" Nov 29 05:17:45 crc kubenswrapper[4799]: I1129 05:17:45.669948 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnnmv"] Nov 29 05:17:45 crc kubenswrapper[4799]: E1129 05:17:45.670674 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41294840-0c22-4c85-ae8f-31f3a00458d4" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Nov 29 05:17:45 crc kubenswrapper[4799]: I1129 05:17:45.670710 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="41294840-0c22-4c85-ae8f-31f3a00458d4" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Nov 29 05:17:45 crc kubenswrapper[4799]: I1129 05:17:45.671072 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="41294840-0c22-4c85-ae8f-31f3a00458d4" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Nov 29 05:17:45 crc kubenswrapper[4799]: I1129 05:17:45.672097 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnnmv" Nov 29 05:17:45 crc kubenswrapper[4799]: I1129 05:17:45.675178 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 29 05:17:45 crc kubenswrapper[4799]: I1129 05:17:45.675479 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ljwtk" Nov 29 05:17:45 crc kubenswrapper[4799]: I1129 05:17:45.675728 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Nov 29 05:17:45 crc kubenswrapper[4799]: I1129 05:17:45.678573 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 29 05:17:45 crc kubenswrapper[4799]: I1129 05:17:45.678618 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 29 05:17:45 crc kubenswrapper[4799]: I1129 05:17:45.678660 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 29 05:17:45 crc kubenswrapper[4799]: I1129 05:17:45.685166 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnnmv"] Nov 29 05:17:45 crc kubenswrapper[4799]: I1129 05:17:45.691594 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9783465-b33b-428e-9691-bc9b341e37ad-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hnnmv\" (UID: \"a9783465-b33b-428e-9691-bc9b341e37ad\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnnmv" Nov 29 05:17:45 crc kubenswrapper[4799]: I1129 05:17:45.691635 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a9783465-b33b-428e-9691-bc9b341e37ad-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hnnmv\" (UID: \"a9783465-b33b-428e-9691-bc9b341e37ad\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnnmv" Nov 29 05:17:45 crc kubenswrapper[4799]: I1129 05:17:45.691669 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crvt4\" (UniqueName: \"kubernetes.io/projected/a9783465-b33b-428e-9691-bc9b341e37ad-kube-api-access-crvt4\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hnnmv\" (UID: \"a9783465-b33b-428e-9691-bc9b341e37ad\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnnmv" Nov 29 05:17:45 crc kubenswrapper[4799]: I1129 05:17:45.691719 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/a9783465-b33b-428e-9691-bc9b341e37ad-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hnnmv\" (UID: \"a9783465-b33b-428e-9691-bc9b341e37ad\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnnmv" Nov 29 05:17:45 crc kubenswrapper[4799]: I1129 05:17:45.691905 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a9783465-b33b-428e-9691-bc9b341e37ad-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hnnmv\" (UID: \"a9783465-b33b-428e-9691-bc9b341e37ad\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnnmv" Nov 29 05:17:45 crc kubenswrapper[4799]: I1129 05:17:45.692145 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a9783465-b33b-428e-9691-bc9b341e37ad-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hnnmv\" (UID: \"a9783465-b33b-428e-9691-bc9b341e37ad\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnnmv" Nov 29 05:17:45 crc kubenswrapper[4799]: I1129 05:17:45.794126 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9783465-b33b-428e-9691-bc9b341e37ad-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hnnmv\" (UID: \"a9783465-b33b-428e-9691-bc9b341e37ad\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnnmv" Nov 29 05:17:45 crc kubenswrapper[4799]: I1129 05:17:45.794184 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a9783465-b33b-428e-9691-bc9b341e37ad-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hnnmv\" (UID: \"a9783465-b33b-428e-9691-bc9b341e37ad\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnnmv" Nov 29 05:17:45 crc kubenswrapper[4799]: I1129 05:17:45.794211 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crvt4\" (UniqueName: \"kubernetes.io/projected/a9783465-b33b-428e-9691-bc9b341e37ad-kube-api-access-crvt4\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hnnmv\" (UID: \"a9783465-b33b-428e-9691-bc9b341e37ad\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnnmv" Nov 29 05:17:45 crc kubenswrapper[4799]: I1129 05:17:45.794266 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/a9783465-b33b-428e-9691-bc9b341e37ad-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hnnmv\" (UID: \"a9783465-b33b-428e-9691-bc9b341e37ad\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnnmv" Nov 29 05:17:45 crc kubenswrapper[4799]: I1129 05:17:45.794356 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a9783465-b33b-428e-9691-bc9b341e37ad-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hnnmv\" (UID: \"a9783465-b33b-428e-9691-bc9b341e37ad\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnnmv" Nov 29 05:17:45 crc kubenswrapper[4799]: I1129 05:17:45.795343 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a9783465-b33b-428e-9691-bc9b341e37ad-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hnnmv\" (UID: \"a9783465-b33b-428e-9691-bc9b341e37ad\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnnmv" Nov 29 05:17:45 crc kubenswrapper[4799]: I1129 05:17:45.795613 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/a9783465-b33b-428e-9691-bc9b341e37ad-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hnnmv\" (UID: \"a9783465-b33b-428e-9691-bc9b341e37ad\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnnmv" Nov 29 05:17:45 crc kubenswrapper[4799]: I1129 05:17:45.799739 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a9783465-b33b-428e-9691-bc9b341e37ad-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hnnmv\" (UID: \"a9783465-b33b-428e-9691-bc9b341e37ad\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnnmv" Nov 29 05:17:45 crc kubenswrapper[4799]: I1129 05:17:45.799747 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a9783465-b33b-428e-9691-bc9b341e37ad-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hnnmv\" (UID: \"a9783465-b33b-428e-9691-bc9b341e37ad\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnnmv" Nov 29 05:17:45 crc kubenswrapper[4799]: I1129 05:17:45.813435 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a9783465-b33b-428e-9691-bc9b341e37ad-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hnnmv\" (UID: \"a9783465-b33b-428e-9691-bc9b341e37ad\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnnmv" Nov 29 05:17:45 crc kubenswrapper[4799]: I1129 05:17:45.814335 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9783465-b33b-428e-9691-bc9b341e37ad-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hnnmv\" (UID: \"a9783465-b33b-428e-9691-bc9b341e37ad\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnnmv" Nov 29 05:17:45 crc kubenswrapper[4799]: I1129 05:17:45.817385 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crvt4\" (UniqueName: \"kubernetes.io/projected/a9783465-b33b-428e-9691-bc9b341e37ad-kube-api-access-crvt4\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hnnmv\" (UID: \"a9783465-b33b-428e-9691-bc9b341e37ad\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnnmv" Nov 29 05:17:46 crc kubenswrapper[4799]: I1129 05:17:46.000300 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnnmv" Nov 29 05:17:46 crc kubenswrapper[4799]: I1129 05:17:46.626282 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnnmv"] Nov 29 05:17:47 crc kubenswrapper[4799]: I1129 05:17:47.546105 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnnmv" event={"ID":"a9783465-b33b-428e-9691-bc9b341e37ad","Type":"ContainerStarted","Data":"5e3d26536974fbb3979e1e30290c11014c179d6c66318b619f0da1f1cbda60bc"} Nov 29 05:17:47 crc kubenswrapper[4799]: I1129 05:17:47.546539 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnnmv" event={"ID":"a9783465-b33b-428e-9691-bc9b341e37ad","Type":"ContainerStarted","Data":"29613ff75fb032dd0852ac01c1d5dba41729319e1ae06715b9b6f6cd96558d09"} Nov 29 05:17:47 crc kubenswrapper[4799]: I1129 05:17:47.570812 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnnmv" podStartSLOduration=2.087956693 podStartE2EDuration="2.57077412s" podCreationTimestamp="2025-11-29 05:17:45 +0000 UTC" firstStartedPulling="2025-11-29 05:17:46.63713556 +0000 UTC m=+2342.280065960" lastFinishedPulling="2025-11-29 05:17:47.119952987 +0000 UTC m=+2342.762883387" observedRunningTime="2025-11-29 05:17:47.570648887 +0000 UTC m=+2343.213579297" watchObservedRunningTime="2025-11-29 05:17:47.57077412 +0000 UTC m=+2343.213704520" Nov 29 05:17:49 crc kubenswrapper[4799]: I1129 05:17:49.659058 4799 scope.go:117] "RemoveContainer" containerID="d0281928bf1426881365580d5726280995a377a80c626b0654de38a8828878e1" Nov 29 05:17:49 crc kubenswrapper[4799]: E1129 05:17:49.661210 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:18:02 crc kubenswrapper[4799]: I1129 05:18:02.659499 4799 scope.go:117] "RemoveContainer" containerID="d0281928bf1426881365580d5726280995a377a80c626b0654de38a8828878e1" Nov 29 05:18:02 crc kubenswrapper[4799]: E1129 05:18:02.660274 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:18:17 crc kubenswrapper[4799]: I1129 05:18:17.658814 4799 scope.go:117] "RemoveContainer" containerID="d0281928bf1426881365580d5726280995a377a80c626b0654de38a8828878e1" Nov 29 05:18:17 crc kubenswrapper[4799]: E1129 05:18:17.659905 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:18:28 crc kubenswrapper[4799]: I1129 05:18:28.659543 4799 scope.go:117] "RemoveContainer" containerID="d0281928bf1426881365580d5726280995a377a80c626b0654de38a8828878e1" Nov 29 05:18:28 crc kubenswrapper[4799]: E1129 05:18:28.660625 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:18:43 crc kubenswrapper[4799]: I1129 05:18:43.659224 4799 scope.go:117] "RemoveContainer" containerID="d0281928bf1426881365580d5726280995a377a80c626b0654de38a8828878e1" Nov 29 05:18:43 crc kubenswrapper[4799]: E1129 05:18:43.663048 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:18:58 crc kubenswrapper[4799]: I1129 05:18:58.660280 4799 scope.go:117] "RemoveContainer" containerID="d0281928bf1426881365580d5726280995a377a80c626b0654de38a8828878e1" Nov 29 05:18:58 crc kubenswrapper[4799]: E1129 05:18:58.661658 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:19:04 crc kubenswrapper[4799]: I1129 05:19:04.464028 4799 generic.go:334] "Generic (PLEG): container finished" podID="a9783465-b33b-428e-9691-bc9b341e37ad" containerID="5e3d26536974fbb3979e1e30290c11014c179d6c66318b619f0da1f1cbda60bc" exitCode=0 Nov 29 05:19:04 crc kubenswrapper[4799]: I1129 05:19:04.464074 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnnmv" event={"ID":"a9783465-b33b-428e-9691-bc9b341e37ad","Type":"ContainerDied","Data":"5e3d26536974fbb3979e1e30290c11014c179d6c66318b619f0da1f1cbda60bc"} Nov 29 05:19:05 crc kubenswrapper[4799]: I1129 05:19:05.999057 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnnmv" Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.049531 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a9783465-b33b-428e-9691-bc9b341e37ad-ssh-key\") pod \"a9783465-b33b-428e-9691-bc9b341e37ad\" (UID: \"a9783465-b33b-428e-9691-bc9b341e37ad\") " Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.049606 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9783465-b33b-428e-9691-bc9b341e37ad-ovn-combined-ca-bundle\") pod \"a9783465-b33b-428e-9691-bc9b341e37ad\" (UID: \"a9783465-b33b-428e-9691-bc9b341e37ad\") " Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.049671 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crvt4\" (UniqueName: \"kubernetes.io/projected/a9783465-b33b-428e-9691-bc9b341e37ad-kube-api-access-crvt4\") pod \"a9783465-b33b-428e-9691-bc9b341e37ad\" (UID: \"a9783465-b33b-428e-9691-bc9b341e37ad\") " Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.049842 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a9783465-b33b-428e-9691-bc9b341e37ad-inventory\") pod \"a9783465-b33b-428e-9691-bc9b341e37ad\" (UID: \"a9783465-b33b-428e-9691-bc9b341e37ad\") " Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.049879 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a9783465-b33b-428e-9691-bc9b341e37ad-ceph\") pod \"a9783465-b33b-428e-9691-bc9b341e37ad\" (UID: \"a9783465-b33b-428e-9691-bc9b341e37ad\") " Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.049999 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/a9783465-b33b-428e-9691-bc9b341e37ad-ovncontroller-config-0\") pod \"a9783465-b33b-428e-9691-bc9b341e37ad\" (UID: \"a9783465-b33b-428e-9691-bc9b341e37ad\") " Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.058714 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9783465-b33b-428e-9691-bc9b341e37ad-ceph" (OuterVolumeSpecName: "ceph") pod "a9783465-b33b-428e-9691-bc9b341e37ad" (UID: "a9783465-b33b-428e-9691-bc9b341e37ad"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.061290 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9783465-b33b-428e-9691-bc9b341e37ad-kube-api-access-crvt4" (OuterVolumeSpecName: "kube-api-access-crvt4") pod "a9783465-b33b-428e-9691-bc9b341e37ad" (UID: "a9783465-b33b-428e-9691-bc9b341e37ad"). InnerVolumeSpecName "kube-api-access-crvt4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.063111 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9783465-b33b-428e-9691-bc9b341e37ad-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "a9783465-b33b-428e-9691-bc9b341e37ad" (UID: "a9783465-b33b-428e-9691-bc9b341e37ad"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.084440 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9783465-b33b-428e-9691-bc9b341e37ad-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a9783465-b33b-428e-9691-bc9b341e37ad" (UID: "a9783465-b33b-428e-9691-bc9b341e37ad"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.085107 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9783465-b33b-428e-9691-bc9b341e37ad-inventory" (OuterVolumeSpecName: "inventory") pod "a9783465-b33b-428e-9691-bc9b341e37ad" (UID: "a9783465-b33b-428e-9691-bc9b341e37ad"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.088729 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9783465-b33b-428e-9691-bc9b341e37ad-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "a9783465-b33b-428e-9691-bc9b341e37ad" (UID: "a9783465-b33b-428e-9691-bc9b341e37ad"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.151193 4799 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9783465-b33b-428e-9691-bc9b341e37ad-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.151618 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crvt4\" (UniqueName: \"kubernetes.io/projected/a9783465-b33b-428e-9691-bc9b341e37ad-kube-api-access-crvt4\") on node \"crc\" DevicePath \"\"" Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.151633 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a9783465-b33b-428e-9691-bc9b341e37ad-inventory\") on node \"crc\" DevicePath \"\"" Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.151647 4799 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a9783465-b33b-428e-9691-bc9b341e37ad-ceph\") on node \"crc\" DevicePath \"\"" Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.151663 4799 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/a9783465-b33b-428e-9691-bc9b341e37ad-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.151675 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a9783465-b33b-428e-9691-bc9b341e37ad-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.491219 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnnmv" event={"ID":"a9783465-b33b-428e-9691-bc9b341e37ad","Type":"ContainerDied","Data":"29613ff75fb032dd0852ac01c1d5dba41729319e1ae06715b9b6f6cd96558d09"} Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.491283 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="29613ff75fb032dd0852ac01c1d5dba41729319e1ae06715b9b6f6cd96558d09" Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.491348 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnnmv" Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.642312 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9"] Nov 29 05:19:06 crc kubenswrapper[4799]: E1129 05:19:06.643253 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9783465-b33b-428e-9691-bc9b341e37ad" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.643289 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9783465-b33b-428e-9691-bc9b341e37ad" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.643658 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9783465-b33b-428e-9691-bc9b341e37ad" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.644854 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9" Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.648549 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.649086 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.649739 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.649827 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ljwtk" Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.649875 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.653561 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.656116 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.678622 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9"] Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.772141 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9\" (UID: \"8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9" Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.772275 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9\" (UID: \"8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9" Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.772327 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9\" (UID: \"8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9" Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.773720 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9\" (UID: \"8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9" Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.773782 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9\" (UID: \"8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9" Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.773882 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9\" (UID: \"8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9" Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.773993 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-427zq\" (UniqueName: \"kubernetes.io/projected/8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34-kube-api-access-427zq\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9\" (UID: \"8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9" Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.876871 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9\" (UID: \"8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9" Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.876956 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9\" (UID: \"8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9" Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.877030 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9\" (UID: \"8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9" Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.877145 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-427zq\" (UniqueName: \"kubernetes.io/projected/8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34-kube-api-access-427zq\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9\" (UID: \"8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9" Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.877243 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9\" (UID: \"8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9" Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.877328 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9\" (UID: \"8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9" Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.877405 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9\" (UID: \"8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9" Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.882733 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9\" (UID: \"8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9" Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.883221 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9\" (UID: \"8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9" Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.885026 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9\" (UID: \"8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9" Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.885737 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9\" (UID: \"8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9" Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.887586 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9\" (UID: \"8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9" Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.896725 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9\" (UID: \"8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9" Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.900064 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-427zq\" (UniqueName: \"kubernetes.io/projected/8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34-kube-api-access-427zq\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9\" (UID: \"8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9" Nov 29 05:19:06 crc kubenswrapper[4799]: I1129 05:19:06.990576 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9" Nov 29 05:19:07 crc kubenswrapper[4799]: I1129 05:19:07.599947 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9"] Nov 29 05:19:07 crc kubenswrapper[4799]: W1129 05:19:07.608672 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8f234eaf_4d04_4d43_ab4b_aba8a4ab1e34.slice/crio-885fe59d81e200b0d39a6f17f024b67f60e3277c90c4f9d7cd00d12cf38e96d1 WatchSource:0}: Error finding container 885fe59d81e200b0d39a6f17f024b67f60e3277c90c4f9d7cd00d12cf38e96d1: Status 404 returned error can't find the container with id 885fe59d81e200b0d39a6f17f024b67f60e3277c90c4f9d7cd00d12cf38e96d1 Nov 29 05:19:08 crc kubenswrapper[4799]: I1129 05:19:08.517642 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9" event={"ID":"8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34","Type":"ContainerStarted","Data":"885fe59d81e200b0d39a6f17f024b67f60e3277c90c4f9d7cd00d12cf38e96d1"} Nov 29 05:19:09 crc kubenswrapper[4799]: I1129 05:19:09.533298 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9" event={"ID":"8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34","Type":"ContainerStarted","Data":"81340e8581abd34fe0e94d85d6f170c750ee610fafb7122a28f26541a4623b18"} Nov 29 05:19:09 crc kubenswrapper[4799]: I1129 05:19:09.562902 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9" podStartSLOduration=2.904602388 podStartE2EDuration="3.56287235s" podCreationTimestamp="2025-11-29 05:19:06 +0000 UTC" firstStartedPulling="2025-11-29 05:19:07.612493064 +0000 UTC m=+2423.255423464" lastFinishedPulling="2025-11-29 05:19:08.270762996 +0000 UTC m=+2423.913693426" observedRunningTime="2025-11-29 05:19:09.557563099 +0000 UTC m=+2425.200493499" watchObservedRunningTime="2025-11-29 05:19:09.56287235 +0000 UTC m=+2425.205802790" Nov 29 05:19:11 crc kubenswrapper[4799]: I1129 05:19:11.659263 4799 scope.go:117] "RemoveContainer" containerID="d0281928bf1426881365580d5726280995a377a80c626b0654de38a8828878e1" Nov 29 05:19:11 crc kubenswrapper[4799]: E1129 05:19:11.659756 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:19:23 crc kubenswrapper[4799]: I1129 05:19:23.659701 4799 scope.go:117] "RemoveContainer" containerID="d0281928bf1426881365580d5726280995a377a80c626b0654de38a8828878e1" Nov 29 05:19:23 crc kubenswrapper[4799]: E1129 05:19:23.661650 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:19:35 crc kubenswrapper[4799]: I1129 05:19:35.658807 4799 scope.go:117] "RemoveContainer" containerID="d0281928bf1426881365580d5726280995a377a80c626b0654de38a8828878e1" Nov 29 05:19:35 crc kubenswrapper[4799]: E1129 05:19:35.659692 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:19:48 crc kubenswrapper[4799]: I1129 05:19:48.659617 4799 scope.go:117] "RemoveContainer" containerID="d0281928bf1426881365580d5726280995a377a80c626b0654de38a8828878e1" Nov 29 05:19:48 crc kubenswrapper[4799]: E1129 05:19:48.660563 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:20:00 crc kubenswrapper[4799]: I1129 05:20:00.659276 4799 scope.go:117] "RemoveContainer" containerID="d0281928bf1426881365580d5726280995a377a80c626b0654de38a8828878e1" Nov 29 05:20:00 crc kubenswrapper[4799]: E1129 05:20:00.659882 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:20:13 crc kubenswrapper[4799]: I1129 05:20:13.285466 4799 generic.go:334] "Generic (PLEG): container finished" podID="8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34" containerID="81340e8581abd34fe0e94d85d6f170c750ee610fafb7122a28f26541a4623b18" exitCode=0 Nov 29 05:20:13 crc kubenswrapper[4799]: I1129 05:20:13.285577 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9" event={"ID":"8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34","Type":"ContainerDied","Data":"81340e8581abd34fe0e94d85d6f170c750ee610fafb7122a28f26541a4623b18"} Nov 29 05:20:14 crc kubenswrapper[4799]: I1129 05:20:14.763524 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9" Nov 29 05:20:14 crc kubenswrapper[4799]: I1129 05:20:14.879632 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34-inventory\") pod \"8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34\" (UID: \"8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34\") " Nov 29 05:20:14 crc kubenswrapper[4799]: I1129 05:20:14.880279 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34-ceph\") pod \"8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34\" (UID: \"8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34\") " Nov 29 05:20:14 crc kubenswrapper[4799]: I1129 05:20:14.880493 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-427zq\" (UniqueName: \"kubernetes.io/projected/8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34-kube-api-access-427zq\") pod \"8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34\" (UID: \"8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34\") " Nov 29 05:20:14 crc kubenswrapper[4799]: I1129 05:20:14.880663 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34-neutron-metadata-combined-ca-bundle\") pod \"8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34\" (UID: \"8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34\") " Nov 29 05:20:14 crc kubenswrapper[4799]: I1129 05:20:14.880706 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34-nova-metadata-neutron-config-0\") pod \"8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34\" (UID: \"8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34\") " Nov 29 05:20:14 crc kubenswrapper[4799]: I1129 05:20:14.880729 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34-ssh-key\") pod \"8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34\" (UID: \"8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34\") " Nov 29 05:20:14 crc kubenswrapper[4799]: I1129 05:20:14.880778 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34-neutron-ovn-metadata-agent-neutron-config-0\") pod \"8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34\" (UID: \"8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34\") " Nov 29 05:20:14 crc kubenswrapper[4799]: I1129 05:20:14.886809 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34" (UID: "8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:20:14 crc kubenswrapper[4799]: I1129 05:20:14.887330 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34-ceph" (OuterVolumeSpecName: "ceph") pod "8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34" (UID: "8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:20:14 crc kubenswrapper[4799]: I1129 05:20:14.897164 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34-kube-api-access-427zq" (OuterVolumeSpecName: "kube-api-access-427zq") pod "8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34" (UID: "8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34"). InnerVolumeSpecName "kube-api-access-427zq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:20:14 crc kubenswrapper[4799]: I1129 05:20:14.909906 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34" (UID: "8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:20:14 crc kubenswrapper[4799]: I1129 05:20:14.911596 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34" (UID: "8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:20:14 crc kubenswrapper[4799]: I1129 05:20:14.912952 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34-inventory" (OuterVolumeSpecName: "inventory") pod "8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34" (UID: "8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:20:14 crc kubenswrapper[4799]: I1129 05:20:14.935399 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34" (UID: "8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:20:14 crc kubenswrapper[4799]: I1129 05:20:14.983057 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-427zq\" (UniqueName: \"kubernetes.io/projected/8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34-kube-api-access-427zq\") on node \"crc\" DevicePath \"\"" Nov 29 05:20:14 crc kubenswrapper[4799]: I1129 05:20:14.983094 4799 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 05:20:14 crc kubenswrapper[4799]: I1129 05:20:14.983108 4799 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 29 05:20:14 crc kubenswrapper[4799]: I1129 05:20:14.983118 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 29 05:20:14 crc kubenswrapper[4799]: I1129 05:20:14.983129 4799 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 29 05:20:14 crc kubenswrapper[4799]: I1129 05:20:14.983141 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34-inventory\") on node \"crc\" DevicePath \"\"" Nov 29 05:20:14 crc kubenswrapper[4799]: I1129 05:20:14.983149 4799 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34-ceph\") on node \"crc\" DevicePath \"\"" Nov 29 05:20:15 crc kubenswrapper[4799]: I1129 05:20:15.311407 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9" event={"ID":"8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34","Type":"ContainerDied","Data":"885fe59d81e200b0d39a6f17f024b67f60e3277c90c4f9d7cd00d12cf38e96d1"} Nov 29 05:20:15 crc kubenswrapper[4799]: I1129 05:20:15.311737 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="885fe59d81e200b0d39a6f17f024b67f60e3277c90c4f9d7cd00d12cf38e96d1" Nov 29 05:20:15 crc kubenswrapper[4799]: I1129 05:20:15.311622 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9" Nov 29 05:20:15 crc kubenswrapper[4799]: I1129 05:20:15.485483 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v72s9"] Nov 29 05:20:15 crc kubenswrapper[4799]: E1129 05:20:15.486261 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 29 05:20:15 crc kubenswrapper[4799]: I1129 05:20:15.486339 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 29 05:20:15 crc kubenswrapper[4799]: I1129 05:20:15.486614 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 29 05:20:15 crc kubenswrapper[4799]: I1129 05:20:15.487423 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v72s9" Nov 29 05:20:15 crc kubenswrapper[4799]: I1129 05:20:15.490227 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ljwtk" Nov 29 05:20:15 crc kubenswrapper[4799]: I1129 05:20:15.490351 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 29 05:20:15 crc kubenswrapper[4799]: I1129 05:20:15.493951 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 29 05:20:15 crc kubenswrapper[4799]: I1129 05:20:15.494267 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 29 05:20:15 crc kubenswrapper[4799]: I1129 05:20:15.494320 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 29 05:20:15 crc kubenswrapper[4799]: I1129 05:20:15.495020 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Nov 29 05:20:15 crc kubenswrapper[4799]: I1129 05:20:15.502419 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v72s9"] Nov 29 05:20:15 crc kubenswrapper[4799]: I1129 05:20:15.596071 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/168cfe40-080e-44fa-863e-8eb8b5b5923c-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v72s9\" (UID: \"168cfe40-080e-44fa-863e-8eb8b5b5923c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v72s9" Nov 29 05:20:15 crc kubenswrapper[4799]: I1129 05:20:15.596273 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/168cfe40-080e-44fa-863e-8eb8b5b5923c-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v72s9\" (UID: \"168cfe40-080e-44fa-863e-8eb8b5b5923c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v72s9" Nov 29 05:20:15 crc kubenswrapper[4799]: I1129 05:20:15.596654 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/168cfe40-080e-44fa-863e-8eb8b5b5923c-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v72s9\" (UID: \"168cfe40-080e-44fa-863e-8eb8b5b5923c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v72s9" Nov 29 05:20:15 crc kubenswrapper[4799]: I1129 05:20:15.596705 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/168cfe40-080e-44fa-863e-8eb8b5b5923c-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v72s9\" (UID: \"168cfe40-080e-44fa-863e-8eb8b5b5923c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v72s9" Nov 29 05:20:15 crc kubenswrapper[4799]: I1129 05:20:15.596737 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfvkd\" (UniqueName: \"kubernetes.io/projected/168cfe40-080e-44fa-863e-8eb8b5b5923c-kube-api-access-pfvkd\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v72s9\" (UID: \"168cfe40-080e-44fa-863e-8eb8b5b5923c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v72s9" Nov 29 05:20:15 crc kubenswrapper[4799]: I1129 05:20:15.597924 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/168cfe40-080e-44fa-863e-8eb8b5b5923c-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v72s9\" (UID: \"168cfe40-080e-44fa-863e-8eb8b5b5923c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v72s9" Nov 29 05:20:15 crc kubenswrapper[4799]: I1129 05:20:15.659383 4799 scope.go:117] "RemoveContainer" containerID="d0281928bf1426881365580d5726280995a377a80c626b0654de38a8828878e1" Nov 29 05:20:15 crc kubenswrapper[4799]: E1129 05:20:15.659963 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:20:15 crc kubenswrapper[4799]: I1129 05:20:15.700164 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/168cfe40-080e-44fa-863e-8eb8b5b5923c-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v72s9\" (UID: \"168cfe40-080e-44fa-863e-8eb8b5b5923c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v72s9" Nov 29 05:20:15 crc kubenswrapper[4799]: I1129 05:20:15.700375 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/168cfe40-080e-44fa-863e-8eb8b5b5923c-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v72s9\" (UID: \"168cfe40-080e-44fa-863e-8eb8b5b5923c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v72s9" Nov 29 05:20:15 crc kubenswrapper[4799]: I1129 05:20:15.700501 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/168cfe40-080e-44fa-863e-8eb8b5b5923c-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v72s9\" (UID: \"168cfe40-080e-44fa-863e-8eb8b5b5923c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v72s9" Nov 29 05:20:15 crc kubenswrapper[4799]: I1129 05:20:15.700539 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/168cfe40-080e-44fa-863e-8eb8b5b5923c-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v72s9\" (UID: \"168cfe40-080e-44fa-863e-8eb8b5b5923c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v72s9" Nov 29 05:20:15 crc kubenswrapper[4799]: I1129 05:20:15.700564 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfvkd\" (UniqueName: \"kubernetes.io/projected/168cfe40-080e-44fa-863e-8eb8b5b5923c-kube-api-access-pfvkd\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v72s9\" (UID: \"168cfe40-080e-44fa-863e-8eb8b5b5923c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v72s9" Nov 29 05:20:15 crc kubenswrapper[4799]: I1129 05:20:15.700613 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/168cfe40-080e-44fa-863e-8eb8b5b5923c-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v72s9\" (UID: \"168cfe40-080e-44fa-863e-8eb8b5b5923c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v72s9" Nov 29 05:20:15 crc kubenswrapper[4799]: I1129 05:20:15.706025 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/168cfe40-080e-44fa-863e-8eb8b5b5923c-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v72s9\" (UID: \"168cfe40-080e-44fa-863e-8eb8b5b5923c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v72s9" Nov 29 05:20:15 crc kubenswrapper[4799]: I1129 05:20:15.706748 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/168cfe40-080e-44fa-863e-8eb8b5b5923c-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v72s9\" (UID: \"168cfe40-080e-44fa-863e-8eb8b5b5923c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v72s9" Nov 29 05:20:15 crc kubenswrapper[4799]: I1129 05:20:15.708369 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/168cfe40-080e-44fa-863e-8eb8b5b5923c-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v72s9\" (UID: \"168cfe40-080e-44fa-863e-8eb8b5b5923c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v72s9" Nov 29 05:20:15 crc kubenswrapper[4799]: I1129 05:20:15.709222 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/168cfe40-080e-44fa-863e-8eb8b5b5923c-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v72s9\" (UID: \"168cfe40-080e-44fa-863e-8eb8b5b5923c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v72s9" Nov 29 05:20:15 crc kubenswrapper[4799]: I1129 05:20:15.713104 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/168cfe40-080e-44fa-863e-8eb8b5b5923c-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v72s9\" (UID: \"168cfe40-080e-44fa-863e-8eb8b5b5923c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v72s9" Nov 29 05:20:15 crc kubenswrapper[4799]: I1129 05:20:15.722063 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfvkd\" (UniqueName: \"kubernetes.io/projected/168cfe40-080e-44fa-863e-8eb8b5b5923c-kube-api-access-pfvkd\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v72s9\" (UID: \"168cfe40-080e-44fa-863e-8eb8b5b5923c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v72s9" Nov 29 05:20:15 crc kubenswrapper[4799]: I1129 05:20:15.829165 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v72s9" Nov 29 05:20:16 crc kubenswrapper[4799]: I1129 05:20:16.489935 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v72s9"] Nov 29 05:20:17 crc kubenswrapper[4799]: I1129 05:20:17.344574 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v72s9" event={"ID":"168cfe40-080e-44fa-863e-8eb8b5b5923c","Type":"ContainerStarted","Data":"adbd16ef16eff0429ea4bc0776ad5f5b9537ff6b7c04afdee8c03478d07d6ce4"} Nov 29 05:20:18 crc kubenswrapper[4799]: I1129 05:20:18.360221 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v72s9" event={"ID":"168cfe40-080e-44fa-863e-8eb8b5b5923c","Type":"ContainerStarted","Data":"f92777f7de2d0df1270cab73fdbb4cf9607c4973e3c01b45bb463d948b5d5ed3"} Nov 29 05:20:18 crc kubenswrapper[4799]: I1129 05:20:18.416104 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v72s9" podStartSLOduration=2.858408641 podStartE2EDuration="3.41606388s" podCreationTimestamp="2025-11-29 05:20:15 +0000 UTC" firstStartedPulling="2025-11-29 05:20:16.502467327 +0000 UTC m=+2492.145397737" lastFinishedPulling="2025-11-29 05:20:17.060122566 +0000 UTC m=+2492.703052976" observedRunningTime="2025-11-29 05:20:18.393366761 +0000 UTC m=+2494.036297171" watchObservedRunningTime="2025-11-29 05:20:18.41606388 +0000 UTC m=+2494.058994290" Nov 29 05:20:27 crc kubenswrapper[4799]: I1129 05:20:27.659233 4799 scope.go:117] "RemoveContainer" containerID="d0281928bf1426881365580d5726280995a377a80c626b0654de38a8828878e1" Nov 29 05:20:27 crc kubenswrapper[4799]: E1129 05:20:27.660175 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:20:28 crc kubenswrapper[4799]: I1129 05:20:28.058397 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rf2gd"] Nov 29 05:20:28 crc kubenswrapper[4799]: I1129 05:20:28.060868 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rf2gd" Nov 29 05:20:28 crc kubenswrapper[4799]: I1129 05:20:28.079855 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rf2gd"] Nov 29 05:20:28 crc kubenswrapper[4799]: I1129 05:20:28.197818 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03712fa7-0fb7-4432-81ea-f7d55115da1d-utilities\") pod \"redhat-operators-rf2gd\" (UID: \"03712fa7-0fb7-4432-81ea-f7d55115da1d\") " pod="openshift-marketplace/redhat-operators-rf2gd" Nov 29 05:20:28 crc kubenswrapper[4799]: I1129 05:20:28.197982 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcww2\" (UniqueName: \"kubernetes.io/projected/03712fa7-0fb7-4432-81ea-f7d55115da1d-kube-api-access-fcww2\") pod \"redhat-operators-rf2gd\" (UID: \"03712fa7-0fb7-4432-81ea-f7d55115da1d\") " pod="openshift-marketplace/redhat-operators-rf2gd" Nov 29 05:20:28 crc kubenswrapper[4799]: I1129 05:20:28.198038 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03712fa7-0fb7-4432-81ea-f7d55115da1d-catalog-content\") pod \"redhat-operators-rf2gd\" (UID: \"03712fa7-0fb7-4432-81ea-f7d55115da1d\") " pod="openshift-marketplace/redhat-operators-rf2gd" Nov 29 05:20:28 crc kubenswrapper[4799]: I1129 05:20:28.300234 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fcww2\" (UniqueName: \"kubernetes.io/projected/03712fa7-0fb7-4432-81ea-f7d55115da1d-kube-api-access-fcww2\") pod \"redhat-operators-rf2gd\" (UID: \"03712fa7-0fb7-4432-81ea-f7d55115da1d\") " pod="openshift-marketplace/redhat-operators-rf2gd" Nov 29 05:20:28 crc kubenswrapper[4799]: I1129 05:20:28.300346 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03712fa7-0fb7-4432-81ea-f7d55115da1d-catalog-content\") pod \"redhat-operators-rf2gd\" (UID: \"03712fa7-0fb7-4432-81ea-f7d55115da1d\") " pod="openshift-marketplace/redhat-operators-rf2gd" Nov 29 05:20:28 crc kubenswrapper[4799]: I1129 05:20:28.300405 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03712fa7-0fb7-4432-81ea-f7d55115da1d-utilities\") pod \"redhat-operators-rf2gd\" (UID: \"03712fa7-0fb7-4432-81ea-f7d55115da1d\") " pod="openshift-marketplace/redhat-operators-rf2gd" Nov 29 05:20:28 crc kubenswrapper[4799]: I1129 05:20:28.301192 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03712fa7-0fb7-4432-81ea-f7d55115da1d-catalog-content\") pod \"redhat-operators-rf2gd\" (UID: \"03712fa7-0fb7-4432-81ea-f7d55115da1d\") " pod="openshift-marketplace/redhat-operators-rf2gd" Nov 29 05:20:28 crc kubenswrapper[4799]: I1129 05:20:28.301224 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03712fa7-0fb7-4432-81ea-f7d55115da1d-utilities\") pod \"redhat-operators-rf2gd\" (UID: \"03712fa7-0fb7-4432-81ea-f7d55115da1d\") " pod="openshift-marketplace/redhat-operators-rf2gd" Nov 29 05:20:28 crc kubenswrapper[4799]: I1129 05:20:28.340436 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fcww2\" (UniqueName: \"kubernetes.io/projected/03712fa7-0fb7-4432-81ea-f7d55115da1d-kube-api-access-fcww2\") pod \"redhat-operators-rf2gd\" (UID: \"03712fa7-0fb7-4432-81ea-f7d55115da1d\") " pod="openshift-marketplace/redhat-operators-rf2gd" Nov 29 05:20:28 crc kubenswrapper[4799]: I1129 05:20:28.380545 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rf2gd" Nov 29 05:20:28 crc kubenswrapper[4799]: I1129 05:20:28.883394 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rf2gd"] Nov 29 05:20:29 crc kubenswrapper[4799]: I1129 05:20:29.491346 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rf2gd" event={"ID":"03712fa7-0fb7-4432-81ea-f7d55115da1d","Type":"ContainerDied","Data":"7678c3416f32c22baa36a2ee63b5a71f1dd62ca89ff170a8349f85c1f6f39f27"} Nov 29 05:20:29 crc kubenswrapper[4799]: I1129 05:20:29.491177 4799 generic.go:334] "Generic (PLEG): container finished" podID="03712fa7-0fb7-4432-81ea-f7d55115da1d" containerID="7678c3416f32c22baa36a2ee63b5a71f1dd62ca89ff170a8349f85c1f6f39f27" exitCode=0 Nov 29 05:20:29 crc kubenswrapper[4799]: I1129 05:20:29.492499 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rf2gd" event={"ID":"03712fa7-0fb7-4432-81ea-f7d55115da1d","Type":"ContainerStarted","Data":"fdd78f5972d5a50eba8934cadb52e4e795f96cbf081f72b88486c692b04b5494"} Nov 29 05:20:30 crc kubenswrapper[4799]: I1129 05:20:30.458765 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hnllr"] Nov 29 05:20:30 crc kubenswrapper[4799]: I1129 05:20:30.463057 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hnllr" Nov 29 05:20:30 crc kubenswrapper[4799]: I1129 05:20:30.497764 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hnllr"] Nov 29 05:20:30 crc kubenswrapper[4799]: I1129 05:20:30.517359 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rf2gd" event={"ID":"03712fa7-0fb7-4432-81ea-f7d55115da1d","Type":"ContainerStarted","Data":"693f136b6b7544a10652b311f3164e55574a314c85fa7f9f3ad8cddbe8276a67"} Nov 29 05:20:30 crc kubenswrapper[4799]: I1129 05:20:30.581869 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c5e5b39-5b47-477c-a347-5587fd1935e6-catalog-content\") pod \"community-operators-hnllr\" (UID: \"4c5e5b39-5b47-477c-a347-5587fd1935e6\") " pod="openshift-marketplace/community-operators-hnllr" Nov 29 05:20:30 crc kubenswrapper[4799]: I1129 05:20:30.582003 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c5e5b39-5b47-477c-a347-5587fd1935e6-utilities\") pod \"community-operators-hnllr\" (UID: \"4c5e5b39-5b47-477c-a347-5587fd1935e6\") " pod="openshift-marketplace/community-operators-hnllr" Nov 29 05:20:30 crc kubenswrapper[4799]: I1129 05:20:30.583208 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4sx5r\" (UniqueName: \"kubernetes.io/projected/4c5e5b39-5b47-477c-a347-5587fd1935e6-kube-api-access-4sx5r\") pod \"community-operators-hnllr\" (UID: \"4c5e5b39-5b47-477c-a347-5587fd1935e6\") " pod="openshift-marketplace/community-operators-hnllr" Nov 29 05:20:30 crc kubenswrapper[4799]: I1129 05:20:30.641296 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-gg6xq"] Nov 29 05:20:30 crc kubenswrapper[4799]: I1129 05:20:30.643674 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gg6xq" Nov 29 05:20:30 crc kubenswrapper[4799]: I1129 05:20:30.677904 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gg6xq"] Nov 29 05:20:30 crc kubenswrapper[4799]: I1129 05:20:30.685549 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c5e5b39-5b47-477c-a347-5587fd1935e6-catalog-content\") pod \"community-operators-hnllr\" (UID: \"4c5e5b39-5b47-477c-a347-5587fd1935e6\") " pod="openshift-marketplace/community-operators-hnllr" Nov 29 05:20:30 crc kubenswrapper[4799]: I1129 05:20:30.685671 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c5e5b39-5b47-477c-a347-5587fd1935e6-utilities\") pod \"community-operators-hnllr\" (UID: \"4c5e5b39-5b47-477c-a347-5587fd1935e6\") " pod="openshift-marketplace/community-operators-hnllr" Nov 29 05:20:30 crc kubenswrapper[4799]: I1129 05:20:30.685733 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4sx5r\" (UniqueName: \"kubernetes.io/projected/4c5e5b39-5b47-477c-a347-5587fd1935e6-kube-api-access-4sx5r\") pod \"community-operators-hnllr\" (UID: \"4c5e5b39-5b47-477c-a347-5587fd1935e6\") " pod="openshift-marketplace/community-operators-hnllr" Nov 29 05:20:30 crc kubenswrapper[4799]: I1129 05:20:30.686613 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c5e5b39-5b47-477c-a347-5587fd1935e6-catalog-content\") pod \"community-operators-hnllr\" (UID: \"4c5e5b39-5b47-477c-a347-5587fd1935e6\") " pod="openshift-marketplace/community-operators-hnllr" Nov 29 05:20:30 crc kubenswrapper[4799]: I1129 05:20:30.687180 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c5e5b39-5b47-477c-a347-5587fd1935e6-utilities\") pod \"community-operators-hnllr\" (UID: \"4c5e5b39-5b47-477c-a347-5587fd1935e6\") " pod="openshift-marketplace/community-operators-hnllr" Nov 29 05:20:30 crc kubenswrapper[4799]: I1129 05:20:30.720285 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4sx5r\" (UniqueName: \"kubernetes.io/projected/4c5e5b39-5b47-477c-a347-5587fd1935e6-kube-api-access-4sx5r\") pod \"community-operators-hnllr\" (UID: \"4c5e5b39-5b47-477c-a347-5587fd1935e6\") " pod="openshift-marketplace/community-operators-hnllr" Nov 29 05:20:30 crc kubenswrapper[4799]: I1129 05:20:30.788040 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57eab018-937e-49d6-b770-ddf49325dc5e-catalog-content\") pod \"certified-operators-gg6xq\" (UID: \"57eab018-937e-49d6-b770-ddf49325dc5e\") " pod="openshift-marketplace/certified-operators-gg6xq" Nov 29 05:20:30 crc kubenswrapper[4799]: I1129 05:20:30.788153 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-425lr\" (UniqueName: \"kubernetes.io/projected/57eab018-937e-49d6-b770-ddf49325dc5e-kube-api-access-425lr\") pod \"certified-operators-gg6xq\" (UID: \"57eab018-937e-49d6-b770-ddf49325dc5e\") " pod="openshift-marketplace/certified-operators-gg6xq" Nov 29 05:20:30 crc kubenswrapper[4799]: I1129 05:20:30.788183 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57eab018-937e-49d6-b770-ddf49325dc5e-utilities\") pod \"certified-operators-gg6xq\" (UID: \"57eab018-937e-49d6-b770-ddf49325dc5e\") " pod="openshift-marketplace/certified-operators-gg6xq" Nov 29 05:20:30 crc kubenswrapper[4799]: I1129 05:20:30.797262 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hnllr" Nov 29 05:20:30 crc kubenswrapper[4799]: I1129 05:20:30.890524 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57eab018-937e-49d6-b770-ddf49325dc5e-catalog-content\") pod \"certified-operators-gg6xq\" (UID: \"57eab018-937e-49d6-b770-ddf49325dc5e\") " pod="openshift-marketplace/certified-operators-gg6xq" Nov 29 05:20:30 crc kubenswrapper[4799]: I1129 05:20:30.890652 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-425lr\" (UniqueName: \"kubernetes.io/projected/57eab018-937e-49d6-b770-ddf49325dc5e-kube-api-access-425lr\") pod \"certified-operators-gg6xq\" (UID: \"57eab018-937e-49d6-b770-ddf49325dc5e\") " pod="openshift-marketplace/certified-operators-gg6xq" Nov 29 05:20:30 crc kubenswrapper[4799]: I1129 05:20:30.890673 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57eab018-937e-49d6-b770-ddf49325dc5e-utilities\") pod \"certified-operators-gg6xq\" (UID: \"57eab018-937e-49d6-b770-ddf49325dc5e\") " pod="openshift-marketplace/certified-operators-gg6xq" Nov 29 05:20:30 crc kubenswrapper[4799]: I1129 05:20:30.891363 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57eab018-937e-49d6-b770-ddf49325dc5e-utilities\") pod \"certified-operators-gg6xq\" (UID: \"57eab018-937e-49d6-b770-ddf49325dc5e\") " pod="openshift-marketplace/certified-operators-gg6xq" Nov 29 05:20:30 crc kubenswrapper[4799]: I1129 05:20:30.891390 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57eab018-937e-49d6-b770-ddf49325dc5e-catalog-content\") pod \"certified-operators-gg6xq\" (UID: \"57eab018-937e-49d6-b770-ddf49325dc5e\") " pod="openshift-marketplace/certified-operators-gg6xq" Nov 29 05:20:30 crc kubenswrapper[4799]: I1129 05:20:30.915303 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-425lr\" (UniqueName: \"kubernetes.io/projected/57eab018-937e-49d6-b770-ddf49325dc5e-kube-api-access-425lr\") pod \"certified-operators-gg6xq\" (UID: \"57eab018-937e-49d6-b770-ddf49325dc5e\") " pod="openshift-marketplace/certified-operators-gg6xq" Nov 29 05:20:30 crc kubenswrapper[4799]: I1129 05:20:30.975690 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gg6xq" Nov 29 05:20:31 crc kubenswrapper[4799]: I1129 05:20:31.396550 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hnllr"] Nov 29 05:20:31 crc kubenswrapper[4799]: I1129 05:20:31.541979 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hnllr" event={"ID":"4c5e5b39-5b47-477c-a347-5587fd1935e6","Type":"ContainerStarted","Data":"d6f7b770e27e29c98a590a707f1828dedecd4084848cbefad62404a707962bcc"} Nov 29 05:20:31 crc kubenswrapper[4799]: I1129 05:20:31.615602 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gg6xq"] Nov 29 05:20:32 crc kubenswrapper[4799]: I1129 05:20:32.558403 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hnllr" event={"ID":"4c5e5b39-5b47-477c-a347-5587fd1935e6","Type":"ContainerStarted","Data":"fcc9a0af46c83dee9fa93ab15997c6e37d4f60463e0d680e8851613545d2d3db"} Nov 29 05:20:32 crc kubenswrapper[4799]: I1129 05:20:32.560544 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gg6xq" event={"ID":"57eab018-937e-49d6-b770-ddf49325dc5e","Type":"ContainerStarted","Data":"b5cfced908c83615b0dc9a3e3923695cd26520de5148916bde7436b918cca6c4"} Nov 29 05:20:32 crc kubenswrapper[4799]: I1129 05:20:32.561068 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gg6xq" event={"ID":"57eab018-937e-49d6-b770-ddf49325dc5e","Type":"ContainerStarted","Data":"49a12f2ad1bca72f43b9c06b7255e368dc9842ea4766654fe345f47460f7f006"} Nov 29 05:20:33 crc kubenswrapper[4799]: I1129 05:20:33.575923 4799 generic.go:334] "Generic (PLEG): container finished" podID="4c5e5b39-5b47-477c-a347-5587fd1935e6" containerID="fcc9a0af46c83dee9fa93ab15997c6e37d4f60463e0d680e8851613545d2d3db" exitCode=0 Nov 29 05:20:33 crc kubenswrapper[4799]: I1129 05:20:33.576035 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hnllr" event={"ID":"4c5e5b39-5b47-477c-a347-5587fd1935e6","Type":"ContainerDied","Data":"fcc9a0af46c83dee9fa93ab15997c6e37d4f60463e0d680e8851613545d2d3db"} Nov 29 05:20:33 crc kubenswrapper[4799]: I1129 05:20:33.578591 4799 generic.go:334] "Generic (PLEG): container finished" podID="03712fa7-0fb7-4432-81ea-f7d55115da1d" containerID="693f136b6b7544a10652b311f3164e55574a314c85fa7f9f3ad8cddbe8276a67" exitCode=0 Nov 29 05:20:33 crc kubenswrapper[4799]: I1129 05:20:33.578660 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rf2gd" event={"ID":"03712fa7-0fb7-4432-81ea-f7d55115da1d","Type":"ContainerDied","Data":"693f136b6b7544a10652b311f3164e55574a314c85fa7f9f3ad8cddbe8276a67"} Nov 29 05:20:34 crc kubenswrapper[4799]: I1129 05:20:34.592327 4799 generic.go:334] "Generic (PLEG): container finished" podID="57eab018-937e-49d6-b770-ddf49325dc5e" containerID="b5cfced908c83615b0dc9a3e3923695cd26520de5148916bde7436b918cca6c4" exitCode=0 Nov 29 05:20:34 crc kubenswrapper[4799]: I1129 05:20:34.592393 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gg6xq" event={"ID":"57eab018-937e-49d6-b770-ddf49325dc5e","Type":"ContainerDied","Data":"b5cfced908c83615b0dc9a3e3923695cd26520de5148916bde7436b918cca6c4"} Nov 29 05:20:34 crc kubenswrapper[4799]: I1129 05:20:34.603360 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rf2gd" event={"ID":"03712fa7-0fb7-4432-81ea-f7d55115da1d","Type":"ContainerStarted","Data":"abf1d63a17b35fc267ed4daaeee66471aa6c6623c2190c22f9d0c12cb6ea88ff"} Nov 29 05:20:34 crc kubenswrapper[4799]: I1129 05:20:34.654357 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rf2gd" podStartSLOduration=2.057598362 podStartE2EDuration="6.654338554s" podCreationTimestamp="2025-11-29 05:20:28 +0000 UTC" firstStartedPulling="2025-11-29 05:20:29.493457664 +0000 UTC m=+2505.136388074" lastFinishedPulling="2025-11-29 05:20:34.090197866 +0000 UTC m=+2509.733128266" observedRunningTime="2025-11-29 05:20:34.64561301 +0000 UTC m=+2510.288543430" watchObservedRunningTime="2025-11-29 05:20:34.654338554 +0000 UTC m=+2510.297268954" Nov 29 05:20:35 crc kubenswrapper[4799]: I1129 05:20:35.630941 4799 generic.go:334] "Generic (PLEG): container finished" podID="4c5e5b39-5b47-477c-a347-5587fd1935e6" containerID="fcd89052cd432a93eb8cfbe9b789740473f6f08939e05a5cb879af2b094a6c63" exitCode=0 Nov 29 05:20:35 crc kubenswrapper[4799]: I1129 05:20:35.631018 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hnllr" event={"ID":"4c5e5b39-5b47-477c-a347-5587fd1935e6","Type":"ContainerDied","Data":"fcd89052cd432a93eb8cfbe9b789740473f6f08939e05a5cb879af2b094a6c63"} Nov 29 05:20:36 crc kubenswrapper[4799]: I1129 05:20:36.649996 4799 generic.go:334] "Generic (PLEG): container finished" podID="57eab018-937e-49d6-b770-ddf49325dc5e" containerID="aa6eb534e648d2af0702028c2e0712d0bd8c5fed753454aed5ff01df6774de87" exitCode=0 Nov 29 05:20:36 crc kubenswrapper[4799]: I1129 05:20:36.650060 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gg6xq" event={"ID":"57eab018-937e-49d6-b770-ddf49325dc5e","Type":"ContainerDied","Data":"aa6eb534e648d2af0702028c2e0712d0bd8c5fed753454aed5ff01df6774de87"} Nov 29 05:20:37 crc kubenswrapper[4799]: I1129 05:20:37.663680 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gg6xq" event={"ID":"57eab018-937e-49d6-b770-ddf49325dc5e","Type":"ContainerStarted","Data":"17afb9ea7f11634d2d0d7c910e27bbcb0e12f36c784ba5797eadc4bc1defa856"} Nov 29 05:20:37 crc kubenswrapper[4799]: I1129 05:20:37.666567 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hnllr" event={"ID":"4c5e5b39-5b47-477c-a347-5587fd1935e6","Type":"ContainerStarted","Data":"732d8185202096aad4a7a313ad7af54bd551242aec4e2674638c62778b109533"} Nov 29 05:20:37 crc kubenswrapper[4799]: I1129 05:20:37.722752 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-gg6xq" podStartSLOduration=5.119421487 podStartE2EDuration="7.722729595s" podCreationTimestamp="2025-11-29 05:20:30 +0000 UTC" firstStartedPulling="2025-11-29 05:20:34.59561751 +0000 UTC m=+2510.238547910" lastFinishedPulling="2025-11-29 05:20:37.198925618 +0000 UTC m=+2512.841856018" observedRunningTime="2025-11-29 05:20:37.686242748 +0000 UTC m=+2513.329173178" watchObservedRunningTime="2025-11-29 05:20:37.722729595 +0000 UTC m=+2513.365660005" Nov 29 05:20:37 crc kubenswrapper[4799]: I1129 05:20:37.728318 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hnllr" podStartSLOduration=4.917026972 podStartE2EDuration="7.728270221s" podCreationTimestamp="2025-11-29 05:20:30 +0000 UTC" firstStartedPulling="2025-11-29 05:20:33.578501697 +0000 UTC m=+2509.221432127" lastFinishedPulling="2025-11-29 05:20:36.389744956 +0000 UTC m=+2512.032675376" observedRunningTime="2025-11-29 05:20:37.715842975 +0000 UTC m=+2513.358773375" watchObservedRunningTime="2025-11-29 05:20:37.728270221 +0000 UTC m=+2513.371200631" Nov 29 05:20:38 crc kubenswrapper[4799]: I1129 05:20:38.380777 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rf2gd" Nov 29 05:20:38 crc kubenswrapper[4799]: I1129 05:20:38.381213 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rf2gd" Nov 29 05:20:39 crc kubenswrapper[4799]: I1129 05:20:39.439739 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rf2gd" podUID="03712fa7-0fb7-4432-81ea-f7d55115da1d" containerName="registry-server" probeResult="failure" output=< Nov 29 05:20:39 crc kubenswrapper[4799]: timeout: failed to connect service ":50051" within 1s Nov 29 05:20:39 crc kubenswrapper[4799]: > Nov 29 05:20:40 crc kubenswrapper[4799]: I1129 05:20:40.659605 4799 scope.go:117] "RemoveContainer" containerID="d0281928bf1426881365580d5726280995a377a80c626b0654de38a8828878e1" Nov 29 05:20:40 crc kubenswrapper[4799]: E1129 05:20:40.659906 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:20:40 crc kubenswrapper[4799]: I1129 05:20:40.797417 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hnllr" Nov 29 05:20:40 crc kubenswrapper[4799]: I1129 05:20:40.797553 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hnllr" Nov 29 05:20:40 crc kubenswrapper[4799]: I1129 05:20:40.876339 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hnllr" Nov 29 05:20:40 crc kubenswrapper[4799]: I1129 05:20:40.977129 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-gg6xq" Nov 29 05:20:40 crc kubenswrapper[4799]: I1129 05:20:40.977218 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-gg6xq" Nov 29 05:20:41 crc kubenswrapper[4799]: I1129 05:20:41.055291 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-gg6xq" Nov 29 05:20:48 crc kubenswrapper[4799]: I1129 05:20:48.451556 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rf2gd" Nov 29 05:20:48 crc kubenswrapper[4799]: I1129 05:20:48.553152 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rf2gd" Nov 29 05:20:48 crc kubenswrapper[4799]: I1129 05:20:48.705544 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rf2gd"] Nov 29 05:20:49 crc kubenswrapper[4799]: I1129 05:20:49.812673 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rf2gd" podUID="03712fa7-0fb7-4432-81ea-f7d55115da1d" containerName="registry-server" containerID="cri-o://abf1d63a17b35fc267ed4daaeee66471aa6c6623c2190c22f9d0c12cb6ea88ff" gracePeriod=2 Nov 29 05:20:50 crc kubenswrapper[4799]: I1129 05:20:50.382198 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rf2gd" Nov 29 05:20:50 crc kubenswrapper[4799]: I1129 05:20:50.510338 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03712fa7-0fb7-4432-81ea-f7d55115da1d-catalog-content\") pod \"03712fa7-0fb7-4432-81ea-f7d55115da1d\" (UID: \"03712fa7-0fb7-4432-81ea-f7d55115da1d\") " Nov 29 05:20:50 crc kubenswrapper[4799]: I1129 05:20:50.510845 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03712fa7-0fb7-4432-81ea-f7d55115da1d-utilities\") pod \"03712fa7-0fb7-4432-81ea-f7d55115da1d\" (UID: \"03712fa7-0fb7-4432-81ea-f7d55115da1d\") " Nov 29 05:20:50 crc kubenswrapper[4799]: I1129 05:20:50.511196 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcww2\" (UniqueName: \"kubernetes.io/projected/03712fa7-0fb7-4432-81ea-f7d55115da1d-kube-api-access-fcww2\") pod \"03712fa7-0fb7-4432-81ea-f7d55115da1d\" (UID: \"03712fa7-0fb7-4432-81ea-f7d55115da1d\") " Nov 29 05:20:50 crc kubenswrapper[4799]: I1129 05:20:50.511773 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03712fa7-0fb7-4432-81ea-f7d55115da1d-utilities" (OuterVolumeSpecName: "utilities") pod "03712fa7-0fb7-4432-81ea-f7d55115da1d" (UID: "03712fa7-0fb7-4432-81ea-f7d55115da1d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:20:50 crc kubenswrapper[4799]: I1129 05:20:50.520617 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03712fa7-0fb7-4432-81ea-f7d55115da1d-kube-api-access-fcww2" (OuterVolumeSpecName: "kube-api-access-fcww2") pod "03712fa7-0fb7-4432-81ea-f7d55115da1d" (UID: "03712fa7-0fb7-4432-81ea-f7d55115da1d"). InnerVolumeSpecName "kube-api-access-fcww2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:20:50 crc kubenswrapper[4799]: I1129 05:20:50.614211 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03712fa7-0fb7-4432-81ea-f7d55115da1d-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 05:20:50 crc kubenswrapper[4799]: I1129 05:20:50.614271 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcww2\" (UniqueName: \"kubernetes.io/projected/03712fa7-0fb7-4432-81ea-f7d55115da1d-kube-api-access-fcww2\") on node \"crc\" DevicePath \"\"" Nov 29 05:20:50 crc kubenswrapper[4799]: I1129 05:20:50.627850 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03712fa7-0fb7-4432-81ea-f7d55115da1d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "03712fa7-0fb7-4432-81ea-f7d55115da1d" (UID: "03712fa7-0fb7-4432-81ea-f7d55115da1d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:20:50 crc kubenswrapper[4799]: I1129 05:20:50.716452 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03712fa7-0fb7-4432-81ea-f7d55115da1d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 05:20:50 crc kubenswrapper[4799]: I1129 05:20:50.827647 4799 generic.go:334] "Generic (PLEG): container finished" podID="03712fa7-0fb7-4432-81ea-f7d55115da1d" containerID="abf1d63a17b35fc267ed4daaeee66471aa6c6623c2190c22f9d0c12cb6ea88ff" exitCode=0 Nov 29 05:20:50 crc kubenswrapper[4799]: I1129 05:20:50.827876 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rf2gd" event={"ID":"03712fa7-0fb7-4432-81ea-f7d55115da1d","Type":"ContainerDied","Data":"abf1d63a17b35fc267ed4daaeee66471aa6c6623c2190c22f9d0c12cb6ea88ff"} Nov 29 05:20:50 crc kubenswrapper[4799]: I1129 05:20:50.827929 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rf2gd" event={"ID":"03712fa7-0fb7-4432-81ea-f7d55115da1d","Type":"ContainerDied","Data":"fdd78f5972d5a50eba8934cadb52e4e795f96cbf081f72b88486c692b04b5494"} Nov 29 05:20:50 crc kubenswrapper[4799]: I1129 05:20:50.827988 4799 scope.go:117] "RemoveContainer" containerID="abf1d63a17b35fc267ed4daaeee66471aa6c6623c2190c22f9d0c12cb6ea88ff" Nov 29 05:20:50 crc kubenswrapper[4799]: I1129 05:20:50.828059 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rf2gd" Nov 29 05:20:50 crc kubenswrapper[4799]: I1129 05:20:50.873328 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hnllr" Nov 29 05:20:50 crc kubenswrapper[4799]: I1129 05:20:50.877014 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rf2gd"] Nov 29 05:20:50 crc kubenswrapper[4799]: I1129 05:20:50.885702 4799 scope.go:117] "RemoveContainer" containerID="693f136b6b7544a10652b311f3164e55574a314c85fa7f9f3ad8cddbe8276a67" Nov 29 05:20:50 crc kubenswrapper[4799]: I1129 05:20:50.890219 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rf2gd"] Nov 29 05:20:50 crc kubenswrapper[4799]: I1129 05:20:50.923312 4799 scope.go:117] "RemoveContainer" containerID="7678c3416f32c22baa36a2ee63b5a71f1dd62ca89ff170a8349f85c1f6f39f27" Nov 29 05:20:50 crc kubenswrapper[4799]: I1129 05:20:50.998557 4799 scope.go:117] "RemoveContainer" containerID="abf1d63a17b35fc267ed4daaeee66471aa6c6623c2190c22f9d0c12cb6ea88ff" Nov 29 05:20:50 crc kubenswrapper[4799]: E1129 05:20:50.999181 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"abf1d63a17b35fc267ed4daaeee66471aa6c6623c2190c22f9d0c12cb6ea88ff\": container with ID starting with abf1d63a17b35fc267ed4daaeee66471aa6c6623c2190c22f9d0c12cb6ea88ff not found: ID does not exist" containerID="abf1d63a17b35fc267ed4daaeee66471aa6c6623c2190c22f9d0c12cb6ea88ff" Nov 29 05:20:50 crc kubenswrapper[4799]: I1129 05:20:50.999258 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abf1d63a17b35fc267ed4daaeee66471aa6c6623c2190c22f9d0c12cb6ea88ff"} err="failed to get container status \"abf1d63a17b35fc267ed4daaeee66471aa6c6623c2190c22f9d0c12cb6ea88ff\": rpc error: code = NotFound desc = could not find container \"abf1d63a17b35fc267ed4daaeee66471aa6c6623c2190c22f9d0c12cb6ea88ff\": container with ID starting with abf1d63a17b35fc267ed4daaeee66471aa6c6623c2190c22f9d0c12cb6ea88ff not found: ID does not exist" Nov 29 05:20:50 crc kubenswrapper[4799]: I1129 05:20:50.999312 4799 scope.go:117] "RemoveContainer" containerID="693f136b6b7544a10652b311f3164e55574a314c85fa7f9f3ad8cddbe8276a67" Nov 29 05:20:50 crc kubenswrapper[4799]: E1129 05:20:50.999780 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"693f136b6b7544a10652b311f3164e55574a314c85fa7f9f3ad8cddbe8276a67\": container with ID starting with 693f136b6b7544a10652b311f3164e55574a314c85fa7f9f3ad8cddbe8276a67 not found: ID does not exist" containerID="693f136b6b7544a10652b311f3164e55574a314c85fa7f9f3ad8cddbe8276a67" Nov 29 05:20:50 crc kubenswrapper[4799]: I1129 05:20:50.999834 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"693f136b6b7544a10652b311f3164e55574a314c85fa7f9f3ad8cddbe8276a67"} err="failed to get container status \"693f136b6b7544a10652b311f3164e55574a314c85fa7f9f3ad8cddbe8276a67\": rpc error: code = NotFound desc = could not find container \"693f136b6b7544a10652b311f3164e55574a314c85fa7f9f3ad8cddbe8276a67\": container with ID starting with 693f136b6b7544a10652b311f3164e55574a314c85fa7f9f3ad8cddbe8276a67 not found: ID does not exist" Nov 29 05:20:50 crc kubenswrapper[4799]: I1129 05:20:50.999869 4799 scope.go:117] "RemoveContainer" containerID="7678c3416f32c22baa36a2ee63b5a71f1dd62ca89ff170a8349f85c1f6f39f27" Nov 29 05:20:51 crc kubenswrapper[4799]: E1129 05:20:51.000078 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7678c3416f32c22baa36a2ee63b5a71f1dd62ca89ff170a8349f85c1f6f39f27\": container with ID starting with 7678c3416f32c22baa36a2ee63b5a71f1dd62ca89ff170a8349f85c1f6f39f27 not found: ID does not exist" containerID="7678c3416f32c22baa36a2ee63b5a71f1dd62ca89ff170a8349f85c1f6f39f27" Nov 29 05:20:51 crc kubenswrapper[4799]: I1129 05:20:51.000106 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7678c3416f32c22baa36a2ee63b5a71f1dd62ca89ff170a8349f85c1f6f39f27"} err="failed to get container status \"7678c3416f32c22baa36a2ee63b5a71f1dd62ca89ff170a8349f85c1f6f39f27\": rpc error: code = NotFound desc = could not find container \"7678c3416f32c22baa36a2ee63b5a71f1dd62ca89ff170a8349f85c1f6f39f27\": container with ID starting with 7678c3416f32c22baa36a2ee63b5a71f1dd62ca89ff170a8349f85c1f6f39f27 not found: ID does not exist" Nov 29 05:20:51 crc kubenswrapper[4799]: I1129 05:20:51.070384 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-gg6xq" Nov 29 05:20:52 crc kubenswrapper[4799]: I1129 05:20:52.681387 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03712fa7-0fb7-4432-81ea-f7d55115da1d" path="/var/lib/kubelet/pods/03712fa7-0fb7-4432-81ea-f7d55115da1d/volumes" Nov 29 05:20:53 crc kubenswrapper[4799]: I1129 05:20:53.308977 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hnllr"] Nov 29 05:20:53 crc kubenswrapper[4799]: I1129 05:20:53.309802 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hnllr" podUID="4c5e5b39-5b47-477c-a347-5587fd1935e6" containerName="registry-server" containerID="cri-o://732d8185202096aad4a7a313ad7af54bd551242aec4e2674638c62778b109533" gracePeriod=2 Nov 29 05:20:53 crc kubenswrapper[4799]: I1129 05:20:53.818661 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hnllr" Nov 29 05:20:53 crc kubenswrapper[4799]: I1129 05:20:53.871241 4799 generic.go:334] "Generic (PLEG): container finished" podID="4c5e5b39-5b47-477c-a347-5587fd1935e6" containerID="732d8185202096aad4a7a313ad7af54bd551242aec4e2674638c62778b109533" exitCode=0 Nov 29 05:20:53 crc kubenswrapper[4799]: I1129 05:20:53.871283 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hnllr" event={"ID":"4c5e5b39-5b47-477c-a347-5587fd1935e6","Type":"ContainerDied","Data":"732d8185202096aad4a7a313ad7af54bd551242aec4e2674638c62778b109533"} Nov 29 05:20:53 crc kubenswrapper[4799]: I1129 05:20:53.871331 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hnllr" event={"ID":"4c5e5b39-5b47-477c-a347-5587fd1935e6","Type":"ContainerDied","Data":"d6f7b770e27e29c98a590a707f1828dedecd4084848cbefad62404a707962bcc"} Nov 29 05:20:53 crc kubenswrapper[4799]: I1129 05:20:53.871354 4799 scope.go:117] "RemoveContainer" containerID="732d8185202096aad4a7a313ad7af54bd551242aec4e2674638c62778b109533" Nov 29 05:20:53 crc kubenswrapper[4799]: I1129 05:20:53.871981 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hnllr" Nov 29 05:20:53 crc kubenswrapper[4799]: I1129 05:20:53.900058 4799 scope.go:117] "RemoveContainer" containerID="fcd89052cd432a93eb8cfbe9b789740473f6f08939e05a5cb879af2b094a6c63" Nov 29 05:20:53 crc kubenswrapper[4799]: I1129 05:20:53.929493 4799 scope.go:117] "RemoveContainer" containerID="fcc9a0af46c83dee9fa93ab15997c6e37d4f60463e0d680e8851613545d2d3db" Nov 29 05:20:53 crc kubenswrapper[4799]: I1129 05:20:53.969023 4799 scope.go:117] "RemoveContainer" containerID="732d8185202096aad4a7a313ad7af54bd551242aec4e2674638c62778b109533" Nov 29 05:20:53 crc kubenswrapper[4799]: E1129 05:20:53.969665 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"732d8185202096aad4a7a313ad7af54bd551242aec4e2674638c62778b109533\": container with ID starting with 732d8185202096aad4a7a313ad7af54bd551242aec4e2674638c62778b109533 not found: ID does not exist" containerID="732d8185202096aad4a7a313ad7af54bd551242aec4e2674638c62778b109533" Nov 29 05:20:53 crc kubenswrapper[4799]: I1129 05:20:53.969708 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"732d8185202096aad4a7a313ad7af54bd551242aec4e2674638c62778b109533"} err="failed to get container status \"732d8185202096aad4a7a313ad7af54bd551242aec4e2674638c62778b109533\": rpc error: code = NotFound desc = could not find container \"732d8185202096aad4a7a313ad7af54bd551242aec4e2674638c62778b109533\": container with ID starting with 732d8185202096aad4a7a313ad7af54bd551242aec4e2674638c62778b109533 not found: ID does not exist" Nov 29 05:20:53 crc kubenswrapper[4799]: I1129 05:20:53.969737 4799 scope.go:117] "RemoveContainer" containerID="fcd89052cd432a93eb8cfbe9b789740473f6f08939e05a5cb879af2b094a6c63" Nov 29 05:20:53 crc kubenswrapper[4799]: E1129 05:20:53.970149 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fcd89052cd432a93eb8cfbe9b789740473f6f08939e05a5cb879af2b094a6c63\": container with ID starting with fcd89052cd432a93eb8cfbe9b789740473f6f08939e05a5cb879af2b094a6c63 not found: ID does not exist" containerID="fcd89052cd432a93eb8cfbe9b789740473f6f08939e05a5cb879af2b094a6c63" Nov 29 05:20:53 crc kubenswrapper[4799]: I1129 05:20:53.970174 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fcd89052cd432a93eb8cfbe9b789740473f6f08939e05a5cb879af2b094a6c63"} err="failed to get container status \"fcd89052cd432a93eb8cfbe9b789740473f6f08939e05a5cb879af2b094a6c63\": rpc error: code = NotFound desc = could not find container \"fcd89052cd432a93eb8cfbe9b789740473f6f08939e05a5cb879af2b094a6c63\": container with ID starting with fcd89052cd432a93eb8cfbe9b789740473f6f08939e05a5cb879af2b094a6c63 not found: ID does not exist" Nov 29 05:20:53 crc kubenswrapper[4799]: I1129 05:20:53.970190 4799 scope.go:117] "RemoveContainer" containerID="fcc9a0af46c83dee9fa93ab15997c6e37d4f60463e0d680e8851613545d2d3db" Nov 29 05:20:53 crc kubenswrapper[4799]: E1129 05:20:53.970740 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fcc9a0af46c83dee9fa93ab15997c6e37d4f60463e0d680e8851613545d2d3db\": container with ID starting with fcc9a0af46c83dee9fa93ab15997c6e37d4f60463e0d680e8851613545d2d3db not found: ID does not exist" containerID="fcc9a0af46c83dee9fa93ab15997c6e37d4f60463e0d680e8851613545d2d3db" Nov 29 05:20:53 crc kubenswrapper[4799]: I1129 05:20:53.970764 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fcc9a0af46c83dee9fa93ab15997c6e37d4f60463e0d680e8851613545d2d3db"} err="failed to get container status \"fcc9a0af46c83dee9fa93ab15997c6e37d4f60463e0d680e8851613545d2d3db\": rpc error: code = NotFound desc = could not find container \"fcc9a0af46c83dee9fa93ab15997c6e37d4f60463e0d680e8851613545d2d3db\": container with ID starting with fcc9a0af46c83dee9fa93ab15997c6e37d4f60463e0d680e8851613545d2d3db not found: ID does not exist" Nov 29 05:20:54 crc kubenswrapper[4799]: I1129 05:20:54.012115 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c5e5b39-5b47-477c-a347-5587fd1935e6-utilities\") pod \"4c5e5b39-5b47-477c-a347-5587fd1935e6\" (UID: \"4c5e5b39-5b47-477c-a347-5587fd1935e6\") " Nov 29 05:20:54 crc kubenswrapper[4799]: I1129 05:20:54.012299 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c5e5b39-5b47-477c-a347-5587fd1935e6-catalog-content\") pod \"4c5e5b39-5b47-477c-a347-5587fd1935e6\" (UID: \"4c5e5b39-5b47-477c-a347-5587fd1935e6\") " Nov 29 05:20:54 crc kubenswrapper[4799]: I1129 05:20:54.013385 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4sx5r\" (UniqueName: \"kubernetes.io/projected/4c5e5b39-5b47-477c-a347-5587fd1935e6-kube-api-access-4sx5r\") pod \"4c5e5b39-5b47-477c-a347-5587fd1935e6\" (UID: \"4c5e5b39-5b47-477c-a347-5587fd1935e6\") " Nov 29 05:20:54 crc kubenswrapper[4799]: I1129 05:20:54.013695 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c5e5b39-5b47-477c-a347-5587fd1935e6-utilities" (OuterVolumeSpecName: "utilities") pod "4c5e5b39-5b47-477c-a347-5587fd1935e6" (UID: "4c5e5b39-5b47-477c-a347-5587fd1935e6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:20:54 crc kubenswrapper[4799]: I1129 05:20:54.014816 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c5e5b39-5b47-477c-a347-5587fd1935e6-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 05:20:54 crc kubenswrapper[4799]: I1129 05:20:54.023181 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c5e5b39-5b47-477c-a347-5587fd1935e6-kube-api-access-4sx5r" (OuterVolumeSpecName: "kube-api-access-4sx5r") pod "4c5e5b39-5b47-477c-a347-5587fd1935e6" (UID: "4c5e5b39-5b47-477c-a347-5587fd1935e6"). InnerVolumeSpecName "kube-api-access-4sx5r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:20:54 crc kubenswrapper[4799]: I1129 05:20:54.080704 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c5e5b39-5b47-477c-a347-5587fd1935e6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4c5e5b39-5b47-477c-a347-5587fd1935e6" (UID: "4c5e5b39-5b47-477c-a347-5587fd1935e6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:20:54 crc kubenswrapper[4799]: I1129 05:20:54.117499 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c5e5b39-5b47-477c-a347-5587fd1935e6-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 05:20:54 crc kubenswrapper[4799]: I1129 05:20:54.117560 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4sx5r\" (UniqueName: \"kubernetes.io/projected/4c5e5b39-5b47-477c-a347-5587fd1935e6-kube-api-access-4sx5r\") on node \"crc\" DevicePath \"\"" Nov 29 05:20:54 crc kubenswrapper[4799]: I1129 05:20:54.232588 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hnllr"] Nov 29 05:20:54 crc kubenswrapper[4799]: I1129 05:20:54.246161 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hnllr"] Nov 29 05:20:54 crc kubenswrapper[4799]: I1129 05:20:54.672469 4799 scope.go:117] "RemoveContainer" containerID="d0281928bf1426881365580d5726280995a377a80c626b0654de38a8828878e1" Nov 29 05:20:54 crc kubenswrapper[4799]: E1129 05:20:54.672981 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:20:54 crc kubenswrapper[4799]: I1129 05:20:54.683353 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c5e5b39-5b47-477c-a347-5587fd1935e6" path="/var/lib/kubelet/pods/4c5e5b39-5b47-477c-a347-5587fd1935e6/volumes" Nov 29 05:20:55 crc kubenswrapper[4799]: I1129 05:20:55.109414 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gg6xq"] Nov 29 05:20:55 crc kubenswrapper[4799]: I1129 05:20:55.110291 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-gg6xq" podUID="57eab018-937e-49d6-b770-ddf49325dc5e" containerName="registry-server" containerID="cri-o://17afb9ea7f11634d2d0d7c910e27bbcb0e12f36c784ba5797eadc4bc1defa856" gracePeriod=2 Nov 29 05:20:55 crc kubenswrapper[4799]: I1129 05:20:55.654470 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gg6xq" Nov 29 05:20:55 crc kubenswrapper[4799]: I1129 05:20:55.758198 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57eab018-937e-49d6-b770-ddf49325dc5e-catalog-content\") pod \"57eab018-937e-49d6-b770-ddf49325dc5e\" (UID: \"57eab018-937e-49d6-b770-ddf49325dc5e\") " Nov 29 05:20:55 crc kubenswrapper[4799]: I1129 05:20:55.758849 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57eab018-937e-49d6-b770-ddf49325dc5e-utilities\") pod \"57eab018-937e-49d6-b770-ddf49325dc5e\" (UID: \"57eab018-937e-49d6-b770-ddf49325dc5e\") " Nov 29 05:20:55 crc kubenswrapper[4799]: I1129 05:20:55.759073 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-425lr\" (UniqueName: \"kubernetes.io/projected/57eab018-937e-49d6-b770-ddf49325dc5e-kube-api-access-425lr\") pod \"57eab018-937e-49d6-b770-ddf49325dc5e\" (UID: \"57eab018-937e-49d6-b770-ddf49325dc5e\") " Nov 29 05:20:55 crc kubenswrapper[4799]: I1129 05:20:55.761244 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57eab018-937e-49d6-b770-ddf49325dc5e-utilities" (OuterVolumeSpecName: "utilities") pod "57eab018-937e-49d6-b770-ddf49325dc5e" (UID: "57eab018-937e-49d6-b770-ddf49325dc5e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:20:55 crc kubenswrapper[4799]: I1129 05:20:55.767324 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57eab018-937e-49d6-b770-ddf49325dc5e-kube-api-access-425lr" (OuterVolumeSpecName: "kube-api-access-425lr") pod "57eab018-937e-49d6-b770-ddf49325dc5e" (UID: "57eab018-937e-49d6-b770-ddf49325dc5e"). InnerVolumeSpecName "kube-api-access-425lr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:20:55 crc kubenswrapper[4799]: I1129 05:20:55.818354 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57eab018-937e-49d6-b770-ddf49325dc5e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57eab018-937e-49d6-b770-ddf49325dc5e" (UID: "57eab018-937e-49d6-b770-ddf49325dc5e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:20:55 crc kubenswrapper[4799]: I1129 05:20:55.862883 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-425lr\" (UniqueName: \"kubernetes.io/projected/57eab018-937e-49d6-b770-ddf49325dc5e-kube-api-access-425lr\") on node \"crc\" DevicePath \"\"" Nov 29 05:20:55 crc kubenswrapper[4799]: I1129 05:20:55.862931 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57eab018-937e-49d6-b770-ddf49325dc5e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 05:20:55 crc kubenswrapper[4799]: I1129 05:20:55.862948 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57eab018-937e-49d6-b770-ddf49325dc5e-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 05:20:55 crc kubenswrapper[4799]: I1129 05:20:55.921898 4799 generic.go:334] "Generic (PLEG): container finished" podID="57eab018-937e-49d6-b770-ddf49325dc5e" containerID="17afb9ea7f11634d2d0d7c910e27bbcb0e12f36c784ba5797eadc4bc1defa856" exitCode=0 Nov 29 05:20:55 crc kubenswrapper[4799]: I1129 05:20:55.921957 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gg6xq" event={"ID":"57eab018-937e-49d6-b770-ddf49325dc5e","Type":"ContainerDied","Data":"17afb9ea7f11634d2d0d7c910e27bbcb0e12f36c784ba5797eadc4bc1defa856"} Nov 29 05:20:55 crc kubenswrapper[4799]: I1129 05:20:55.921993 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gg6xq" event={"ID":"57eab018-937e-49d6-b770-ddf49325dc5e","Type":"ContainerDied","Data":"49a12f2ad1bca72f43b9c06b7255e368dc9842ea4766654fe345f47460f7f006"} Nov 29 05:20:55 crc kubenswrapper[4799]: I1129 05:20:55.922013 4799 scope.go:117] "RemoveContainer" containerID="17afb9ea7f11634d2d0d7c910e27bbcb0e12f36c784ba5797eadc4bc1defa856" Nov 29 05:20:55 crc kubenswrapper[4799]: I1129 05:20:55.922015 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gg6xq" Nov 29 05:20:55 crc kubenswrapper[4799]: I1129 05:20:55.950369 4799 scope.go:117] "RemoveContainer" containerID="aa6eb534e648d2af0702028c2e0712d0bd8c5fed753454aed5ff01df6774de87" Nov 29 05:20:55 crc kubenswrapper[4799]: I1129 05:20:55.968317 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gg6xq"] Nov 29 05:20:55 crc kubenswrapper[4799]: I1129 05:20:55.979049 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-gg6xq"] Nov 29 05:20:55 crc kubenswrapper[4799]: I1129 05:20:55.993756 4799 scope.go:117] "RemoveContainer" containerID="b5cfced908c83615b0dc9a3e3923695cd26520de5148916bde7436b918cca6c4" Nov 29 05:20:56 crc kubenswrapper[4799]: I1129 05:20:56.043345 4799 scope.go:117] "RemoveContainer" containerID="17afb9ea7f11634d2d0d7c910e27bbcb0e12f36c784ba5797eadc4bc1defa856" Nov 29 05:20:56 crc kubenswrapper[4799]: E1129 05:20:56.044011 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17afb9ea7f11634d2d0d7c910e27bbcb0e12f36c784ba5797eadc4bc1defa856\": container with ID starting with 17afb9ea7f11634d2d0d7c910e27bbcb0e12f36c784ba5797eadc4bc1defa856 not found: ID does not exist" containerID="17afb9ea7f11634d2d0d7c910e27bbcb0e12f36c784ba5797eadc4bc1defa856" Nov 29 05:20:56 crc kubenswrapper[4799]: I1129 05:20:56.044066 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17afb9ea7f11634d2d0d7c910e27bbcb0e12f36c784ba5797eadc4bc1defa856"} err="failed to get container status \"17afb9ea7f11634d2d0d7c910e27bbcb0e12f36c784ba5797eadc4bc1defa856\": rpc error: code = NotFound desc = could not find container \"17afb9ea7f11634d2d0d7c910e27bbcb0e12f36c784ba5797eadc4bc1defa856\": container with ID starting with 17afb9ea7f11634d2d0d7c910e27bbcb0e12f36c784ba5797eadc4bc1defa856 not found: ID does not exist" Nov 29 05:20:56 crc kubenswrapper[4799]: I1129 05:20:56.044105 4799 scope.go:117] "RemoveContainer" containerID="aa6eb534e648d2af0702028c2e0712d0bd8c5fed753454aed5ff01df6774de87" Nov 29 05:20:56 crc kubenswrapper[4799]: E1129 05:20:56.044513 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa6eb534e648d2af0702028c2e0712d0bd8c5fed753454aed5ff01df6774de87\": container with ID starting with aa6eb534e648d2af0702028c2e0712d0bd8c5fed753454aed5ff01df6774de87 not found: ID does not exist" containerID="aa6eb534e648d2af0702028c2e0712d0bd8c5fed753454aed5ff01df6774de87" Nov 29 05:20:56 crc kubenswrapper[4799]: I1129 05:20:56.044589 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa6eb534e648d2af0702028c2e0712d0bd8c5fed753454aed5ff01df6774de87"} err="failed to get container status \"aa6eb534e648d2af0702028c2e0712d0bd8c5fed753454aed5ff01df6774de87\": rpc error: code = NotFound desc = could not find container \"aa6eb534e648d2af0702028c2e0712d0bd8c5fed753454aed5ff01df6774de87\": container with ID starting with aa6eb534e648d2af0702028c2e0712d0bd8c5fed753454aed5ff01df6774de87 not found: ID does not exist" Nov 29 05:20:56 crc kubenswrapper[4799]: I1129 05:20:56.044651 4799 scope.go:117] "RemoveContainer" containerID="b5cfced908c83615b0dc9a3e3923695cd26520de5148916bde7436b918cca6c4" Nov 29 05:20:56 crc kubenswrapper[4799]: E1129 05:20:56.045040 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5cfced908c83615b0dc9a3e3923695cd26520de5148916bde7436b918cca6c4\": container with ID starting with b5cfced908c83615b0dc9a3e3923695cd26520de5148916bde7436b918cca6c4 not found: ID does not exist" containerID="b5cfced908c83615b0dc9a3e3923695cd26520de5148916bde7436b918cca6c4" Nov 29 05:20:56 crc kubenswrapper[4799]: I1129 05:20:56.045092 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5cfced908c83615b0dc9a3e3923695cd26520de5148916bde7436b918cca6c4"} err="failed to get container status \"b5cfced908c83615b0dc9a3e3923695cd26520de5148916bde7436b918cca6c4\": rpc error: code = NotFound desc = could not find container \"b5cfced908c83615b0dc9a3e3923695cd26520de5148916bde7436b918cca6c4\": container with ID starting with b5cfced908c83615b0dc9a3e3923695cd26520de5148916bde7436b918cca6c4 not found: ID does not exist" Nov 29 05:20:56 crc kubenswrapper[4799]: I1129 05:20:56.670807 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57eab018-937e-49d6-b770-ddf49325dc5e" path="/var/lib/kubelet/pods/57eab018-937e-49d6-b770-ddf49325dc5e/volumes" Nov 29 05:21:09 crc kubenswrapper[4799]: I1129 05:21:09.660551 4799 scope.go:117] "RemoveContainer" containerID="d0281928bf1426881365580d5726280995a377a80c626b0654de38a8828878e1" Nov 29 05:21:10 crc kubenswrapper[4799]: I1129 05:21:10.098664 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" event={"ID":"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8","Type":"ContainerStarted","Data":"4a42289995c12e4c41519c0c91fe9e1ab67dd9e0388b75fa426811cd1920c88e"} Nov 29 05:23:37 crc kubenswrapper[4799]: I1129 05:23:37.498174 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 05:23:37 crc kubenswrapper[4799]: I1129 05:23:37.498885 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 05:24:07 crc kubenswrapper[4799]: I1129 05:24:07.500110 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 05:24:07 crc kubenswrapper[4799]: I1129 05:24:07.501235 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 05:24:37 crc kubenswrapper[4799]: I1129 05:24:37.498158 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 05:24:37 crc kubenswrapper[4799]: I1129 05:24:37.499149 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 05:24:37 crc kubenswrapper[4799]: I1129 05:24:37.499234 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-26t88" Nov 29 05:24:37 crc kubenswrapper[4799]: I1129 05:24:37.500592 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4a42289995c12e4c41519c0c91fe9e1ab67dd9e0388b75fa426811cd1920c88e"} pod="openshift-machine-config-operator/machine-config-daemon-26t88" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 29 05:24:37 crc kubenswrapper[4799]: I1129 05:24:37.500711 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" containerID="cri-o://4a42289995c12e4c41519c0c91fe9e1ab67dd9e0388b75fa426811cd1920c88e" gracePeriod=600 Nov 29 05:24:37 crc kubenswrapper[4799]: I1129 05:24:37.727967 4799 generic.go:334] "Generic (PLEG): container finished" podID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerID="4a42289995c12e4c41519c0c91fe9e1ab67dd9e0388b75fa426811cd1920c88e" exitCode=0 Nov 29 05:24:37 crc kubenswrapper[4799]: I1129 05:24:37.728075 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" event={"ID":"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8","Type":"ContainerDied","Data":"4a42289995c12e4c41519c0c91fe9e1ab67dd9e0388b75fa426811cd1920c88e"} Nov 29 05:24:37 crc kubenswrapper[4799]: I1129 05:24:37.728209 4799 scope.go:117] "RemoveContainer" containerID="d0281928bf1426881365580d5726280995a377a80c626b0654de38a8828878e1" Nov 29 05:24:38 crc kubenswrapper[4799]: I1129 05:24:38.752879 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" event={"ID":"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8","Type":"ContainerStarted","Data":"e6d4b3d1d5b5e1eeb13508d67de3d3c3aac52e76337b55e53793185c0d001d82"} Nov 29 05:25:01 crc kubenswrapper[4799]: I1129 05:25:01.041994 4799 generic.go:334] "Generic (PLEG): container finished" podID="168cfe40-080e-44fa-863e-8eb8b5b5923c" containerID="f92777f7de2d0df1270cab73fdbb4cf9607c4973e3c01b45bb463d948b5d5ed3" exitCode=0 Nov 29 05:25:01 crc kubenswrapper[4799]: I1129 05:25:01.042083 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v72s9" event={"ID":"168cfe40-080e-44fa-863e-8eb8b5b5923c","Type":"ContainerDied","Data":"f92777f7de2d0df1270cab73fdbb4cf9607c4973e3c01b45bb463d948b5d5ed3"} Nov 29 05:25:02 crc kubenswrapper[4799]: I1129 05:25:02.608855 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v72s9" Nov 29 05:25:02 crc kubenswrapper[4799]: I1129 05:25:02.666034 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/168cfe40-080e-44fa-863e-8eb8b5b5923c-libvirt-combined-ca-bundle\") pod \"168cfe40-080e-44fa-863e-8eb8b5b5923c\" (UID: \"168cfe40-080e-44fa-863e-8eb8b5b5923c\") " Nov 29 05:25:02 crc kubenswrapper[4799]: I1129 05:25:02.666115 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/168cfe40-080e-44fa-863e-8eb8b5b5923c-inventory\") pod \"168cfe40-080e-44fa-863e-8eb8b5b5923c\" (UID: \"168cfe40-080e-44fa-863e-8eb8b5b5923c\") " Nov 29 05:25:02 crc kubenswrapper[4799]: I1129 05:25:02.666230 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/168cfe40-080e-44fa-863e-8eb8b5b5923c-ceph\") pod \"168cfe40-080e-44fa-863e-8eb8b5b5923c\" (UID: \"168cfe40-080e-44fa-863e-8eb8b5b5923c\") " Nov 29 05:25:02 crc kubenswrapper[4799]: I1129 05:25:02.666265 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pfvkd\" (UniqueName: \"kubernetes.io/projected/168cfe40-080e-44fa-863e-8eb8b5b5923c-kube-api-access-pfvkd\") pod \"168cfe40-080e-44fa-863e-8eb8b5b5923c\" (UID: \"168cfe40-080e-44fa-863e-8eb8b5b5923c\") " Nov 29 05:25:02 crc kubenswrapper[4799]: I1129 05:25:02.666307 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/168cfe40-080e-44fa-863e-8eb8b5b5923c-libvirt-secret-0\") pod \"168cfe40-080e-44fa-863e-8eb8b5b5923c\" (UID: \"168cfe40-080e-44fa-863e-8eb8b5b5923c\") " Nov 29 05:25:02 crc kubenswrapper[4799]: I1129 05:25:02.666366 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/168cfe40-080e-44fa-863e-8eb8b5b5923c-ssh-key\") pod \"168cfe40-080e-44fa-863e-8eb8b5b5923c\" (UID: \"168cfe40-080e-44fa-863e-8eb8b5b5923c\") " Nov 29 05:25:02 crc kubenswrapper[4799]: I1129 05:25:02.677190 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/168cfe40-080e-44fa-863e-8eb8b5b5923c-ceph" (OuterVolumeSpecName: "ceph") pod "168cfe40-080e-44fa-863e-8eb8b5b5923c" (UID: "168cfe40-080e-44fa-863e-8eb8b5b5923c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:25:02 crc kubenswrapper[4799]: I1129 05:25:02.688424 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/168cfe40-080e-44fa-863e-8eb8b5b5923c-kube-api-access-pfvkd" (OuterVolumeSpecName: "kube-api-access-pfvkd") pod "168cfe40-080e-44fa-863e-8eb8b5b5923c" (UID: "168cfe40-080e-44fa-863e-8eb8b5b5923c"). InnerVolumeSpecName "kube-api-access-pfvkd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:25:02 crc kubenswrapper[4799]: I1129 05:25:02.689712 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/168cfe40-080e-44fa-863e-8eb8b5b5923c-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "168cfe40-080e-44fa-863e-8eb8b5b5923c" (UID: "168cfe40-080e-44fa-863e-8eb8b5b5923c"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:25:02 crc kubenswrapper[4799]: I1129 05:25:02.702445 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/168cfe40-080e-44fa-863e-8eb8b5b5923c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "168cfe40-080e-44fa-863e-8eb8b5b5923c" (UID: "168cfe40-080e-44fa-863e-8eb8b5b5923c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:25:02 crc kubenswrapper[4799]: I1129 05:25:02.710892 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/168cfe40-080e-44fa-863e-8eb8b5b5923c-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "168cfe40-080e-44fa-863e-8eb8b5b5923c" (UID: "168cfe40-080e-44fa-863e-8eb8b5b5923c"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:25:02 crc kubenswrapper[4799]: I1129 05:25:02.730427 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/168cfe40-080e-44fa-863e-8eb8b5b5923c-inventory" (OuterVolumeSpecName: "inventory") pod "168cfe40-080e-44fa-863e-8eb8b5b5923c" (UID: "168cfe40-080e-44fa-863e-8eb8b5b5923c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:25:02 crc kubenswrapper[4799]: I1129 05:25:02.769959 4799 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/168cfe40-080e-44fa-863e-8eb8b5b5923c-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 05:25:02 crc kubenswrapper[4799]: I1129 05:25:02.770282 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/168cfe40-080e-44fa-863e-8eb8b5b5923c-inventory\") on node \"crc\" DevicePath \"\"" Nov 29 05:25:02 crc kubenswrapper[4799]: I1129 05:25:02.770439 4799 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/168cfe40-080e-44fa-863e-8eb8b5b5923c-ceph\") on node \"crc\" DevicePath \"\"" Nov 29 05:25:02 crc kubenswrapper[4799]: I1129 05:25:02.770546 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pfvkd\" (UniqueName: \"kubernetes.io/projected/168cfe40-080e-44fa-863e-8eb8b5b5923c-kube-api-access-pfvkd\") on node \"crc\" DevicePath \"\"" Nov 29 05:25:02 crc kubenswrapper[4799]: I1129 05:25:02.770660 4799 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/168cfe40-080e-44fa-863e-8eb8b5b5923c-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Nov 29 05:25:02 crc kubenswrapper[4799]: I1129 05:25:02.770745 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/168cfe40-080e-44fa-863e-8eb8b5b5923c-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.076334 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v72s9" event={"ID":"168cfe40-080e-44fa-863e-8eb8b5b5923c","Type":"ContainerDied","Data":"adbd16ef16eff0429ea4bc0776ad5f5b9537ff6b7c04afdee8c03478d07d6ce4"} Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.076409 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="adbd16ef16eff0429ea4bc0776ad5f5b9537ff6b7c04afdee8c03478d07d6ce4" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.076465 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v72s9" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.277060 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b"] Nov 29 05:25:03 crc kubenswrapper[4799]: E1129 05:25:03.277740 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="168cfe40-080e-44fa-863e-8eb8b5b5923c" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.277778 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="168cfe40-080e-44fa-863e-8eb8b5b5923c" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 29 05:25:03 crc kubenswrapper[4799]: E1129 05:25:03.277840 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03712fa7-0fb7-4432-81ea-f7d55115da1d" containerName="registry-server" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.277854 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="03712fa7-0fb7-4432-81ea-f7d55115da1d" containerName="registry-server" Nov 29 05:25:03 crc kubenswrapper[4799]: E1129 05:25:03.277882 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03712fa7-0fb7-4432-81ea-f7d55115da1d" containerName="extract-utilities" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.277895 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="03712fa7-0fb7-4432-81ea-f7d55115da1d" containerName="extract-utilities" Nov 29 05:25:03 crc kubenswrapper[4799]: E1129 05:25:03.277922 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c5e5b39-5b47-477c-a347-5587fd1935e6" containerName="extract-utilities" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.277934 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c5e5b39-5b47-477c-a347-5587fd1935e6" containerName="extract-utilities" Nov 29 05:25:03 crc kubenswrapper[4799]: E1129 05:25:03.277955 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c5e5b39-5b47-477c-a347-5587fd1935e6" containerName="extract-content" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.277967 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c5e5b39-5b47-477c-a347-5587fd1935e6" containerName="extract-content" Nov 29 05:25:03 crc kubenswrapper[4799]: E1129 05:25:03.277995 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57eab018-937e-49d6-b770-ddf49325dc5e" containerName="extract-utilities" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.278007 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="57eab018-937e-49d6-b770-ddf49325dc5e" containerName="extract-utilities" Nov 29 05:25:03 crc kubenswrapper[4799]: E1129 05:25:03.278025 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03712fa7-0fb7-4432-81ea-f7d55115da1d" containerName="extract-content" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.278038 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="03712fa7-0fb7-4432-81ea-f7d55115da1d" containerName="extract-content" Nov 29 05:25:03 crc kubenswrapper[4799]: E1129 05:25:03.278063 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57eab018-937e-49d6-b770-ddf49325dc5e" containerName="registry-server" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.278075 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="57eab018-937e-49d6-b770-ddf49325dc5e" containerName="registry-server" Nov 29 05:25:03 crc kubenswrapper[4799]: E1129 05:25:03.278101 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c5e5b39-5b47-477c-a347-5587fd1935e6" containerName="registry-server" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.278115 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c5e5b39-5b47-477c-a347-5587fd1935e6" containerName="registry-server" Nov 29 05:25:03 crc kubenswrapper[4799]: E1129 05:25:03.278137 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57eab018-937e-49d6-b770-ddf49325dc5e" containerName="extract-content" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.278149 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="57eab018-937e-49d6-b770-ddf49325dc5e" containerName="extract-content" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.278564 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="03712fa7-0fb7-4432-81ea-f7d55115da1d" containerName="registry-server" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.278602 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c5e5b39-5b47-477c-a347-5587fd1935e6" containerName="registry-server" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.278644 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="57eab018-937e-49d6-b770-ddf49325dc5e" containerName="registry-server" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.278671 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="168cfe40-080e-44fa-863e-8eb8b5b5923c" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.279968 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.286509 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.286818 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.287279 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ljwtk" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.287348 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.287611 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.287757 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.287967 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.294829 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ceph-nova" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.295191 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.304002 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b"] Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.384669 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/26a0b212-c652-4475-8210-4b62b653ed79-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b\" (UID: \"26a0b212-c652-4475-8210-4b62b653ed79\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.384743 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/26a0b212-c652-4475-8210-4b62b653ed79-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b\" (UID: \"26a0b212-c652-4475-8210-4b62b653ed79\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.384839 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/26a0b212-c652-4475-8210-4b62b653ed79-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b\" (UID: \"26a0b212-c652-4475-8210-4b62b653ed79\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.385290 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/26a0b212-c652-4475-8210-4b62b653ed79-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b\" (UID: \"26a0b212-c652-4475-8210-4b62b653ed79\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.385393 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/26a0b212-c652-4475-8210-4b62b653ed79-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b\" (UID: \"26a0b212-c652-4475-8210-4b62b653ed79\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.385428 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/26a0b212-c652-4475-8210-4b62b653ed79-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b\" (UID: \"26a0b212-c652-4475-8210-4b62b653ed79\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.385491 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/26a0b212-c652-4475-8210-4b62b653ed79-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b\" (UID: \"26a0b212-c652-4475-8210-4b62b653ed79\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.385598 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/26a0b212-c652-4475-8210-4b62b653ed79-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b\" (UID: \"26a0b212-c652-4475-8210-4b62b653ed79\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.385703 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26a0b212-c652-4475-8210-4b62b653ed79-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b\" (UID: \"26a0b212-c652-4475-8210-4b62b653ed79\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.385763 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/26a0b212-c652-4475-8210-4b62b653ed79-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b\" (UID: \"26a0b212-c652-4475-8210-4b62b653ed79\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.385868 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ttvf\" (UniqueName: \"kubernetes.io/projected/26a0b212-c652-4475-8210-4b62b653ed79-kube-api-access-7ttvf\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b\" (UID: \"26a0b212-c652-4475-8210-4b62b653ed79\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.486781 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/26a0b212-c652-4475-8210-4b62b653ed79-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b\" (UID: \"26a0b212-c652-4475-8210-4b62b653ed79\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.486913 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/26a0b212-c652-4475-8210-4b62b653ed79-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b\" (UID: \"26a0b212-c652-4475-8210-4b62b653ed79\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.486955 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/26a0b212-c652-4475-8210-4b62b653ed79-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b\" (UID: \"26a0b212-c652-4475-8210-4b62b653ed79\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.486989 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/26a0b212-c652-4475-8210-4b62b653ed79-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b\" (UID: \"26a0b212-c652-4475-8210-4b62b653ed79\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.487049 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/26a0b212-c652-4475-8210-4b62b653ed79-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b\" (UID: \"26a0b212-c652-4475-8210-4b62b653ed79\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.487107 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26a0b212-c652-4475-8210-4b62b653ed79-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b\" (UID: \"26a0b212-c652-4475-8210-4b62b653ed79\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.487176 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/26a0b212-c652-4475-8210-4b62b653ed79-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b\" (UID: \"26a0b212-c652-4475-8210-4b62b653ed79\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.487226 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ttvf\" (UniqueName: \"kubernetes.io/projected/26a0b212-c652-4475-8210-4b62b653ed79-kube-api-access-7ttvf\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b\" (UID: \"26a0b212-c652-4475-8210-4b62b653ed79\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.487264 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/26a0b212-c652-4475-8210-4b62b653ed79-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b\" (UID: \"26a0b212-c652-4475-8210-4b62b653ed79\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.487298 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/26a0b212-c652-4475-8210-4b62b653ed79-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b\" (UID: \"26a0b212-c652-4475-8210-4b62b653ed79\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.487372 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/26a0b212-c652-4475-8210-4b62b653ed79-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b\" (UID: \"26a0b212-c652-4475-8210-4b62b653ed79\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.488822 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/26a0b212-c652-4475-8210-4b62b653ed79-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b\" (UID: \"26a0b212-c652-4475-8210-4b62b653ed79\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.489836 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/26a0b212-c652-4475-8210-4b62b653ed79-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b\" (UID: \"26a0b212-c652-4475-8210-4b62b653ed79\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.495051 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/26a0b212-c652-4475-8210-4b62b653ed79-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b\" (UID: \"26a0b212-c652-4475-8210-4b62b653ed79\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.496698 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/26a0b212-c652-4475-8210-4b62b653ed79-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b\" (UID: \"26a0b212-c652-4475-8210-4b62b653ed79\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.497273 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/26a0b212-c652-4475-8210-4b62b653ed79-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b\" (UID: \"26a0b212-c652-4475-8210-4b62b653ed79\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.504729 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/26a0b212-c652-4475-8210-4b62b653ed79-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b\" (UID: \"26a0b212-c652-4475-8210-4b62b653ed79\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.507337 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26a0b212-c652-4475-8210-4b62b653ed79-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b\" (UID: \"26a0b212-c652-4475-8210-4b62b653ed79\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.509569 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/26a0b212-c652-4475-8210-4b62b653ed79-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b\" (UID: \"26a0b212-c652-4475-8210-4b62b653ed79\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.510633 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/26a0b212-c652-4475-8210-4b62b653ed79-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b\" (UID: \"26a0b212-c652-4475-8210-4b62b653ed79\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.514944 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/26a0b212-c652-4475-8210-4b62b653ed79-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b\" (UID: \"26a0b212-c652-4475-8210-4b62b653ed79\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.517618 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ttvf\" (UniqueName: \"kubernetes.io/projected/26a0b212-c652-4475-8210-4b62b653ed79-kube-api-access-7ttvf\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b\" (UID: \"26a0b212-c652-4475-8210-4b62b653ed79\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b" Nov 29 05:25:03 crc kubenswrapper[4799]: I1129 05:25:03.609616 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b" Nov 29 05:25:04 crc kubenswrapper[4799]: I1129 05:25:04.282300 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b"] Nov 29 05:25:04 crc kubenswrapper[4799]: I1129 05:25:04.302500 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 29 05:25:05 crc kubenswrapper[4799]: I1129 05:25:05.118913 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b" event={"ID":"26a0b212-c652-4475-8210-4b62b653ed79","Type":"ContainerStarted","Data":"afe9ac99e8c014634b44f791e0d91e88a2f77af7e01c558b56355a9116f38e4b"} Nov 29 05:25:06 crc kubenswrapper[4799]: I1129 05:25:06.136508 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b" event={"ID":"26a0b212-c652-4475-8210-4b62b653ed79","Type":"ContainerStarted","Data":"94d4a5e46c1fefdf618e24b6387dae4a821483a636fec1ff9ea7faeb4dbf2b69"} Nov 29 05:25:06 crc kubenswrapper[4799]: I1129 05:25:06.163837 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b" podStartSLOduration=2.48314969 podStartE2EDuration="3.163786733s" podCreationTimestamp="2025-11-29 05:25:03 +0000 UTC" firstStartedPulling="2025-11-29 05:25:04.302078477 +0000 UTC m=+2779.945008917" lastFinishedPulling="2025-11-29 05:25:04.98271555 +0000 UTC m=+2780.625645960" observedRunningTime="2025-11-29 05:25:06.16322342 +0000 UTC m=+2781.806153860" watchObservedRunningTime="2025-11-29 05:25:06.163786733 +0000 UTC m=+2781.806717143" Nov 29 05:25:55 crc kubenswrapper[4799]: I1129 05:25:55.524594 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-p8w8l"] Nov 29 05:25:55 crc kubenswrapper[4799]: I1129 05:25:55.529666 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p8w8l" Nov 29 05:25:55 crc kubenswrapper[4799]: I1129 05:25:55.548264 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-p8w8l"] Nov 29 05:25:55 crc kubenswrapper[4799]: I1129 05:25:55.664063 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1e5797a-9ff4-466b-83c7-325510a31cf1-catalog-content\") pod \"redhat-marketplace-p8w8l\" (UID: \"c1e5797a-9ff4-466b-83c7-325510a31cf1\") " pod="openshift-marketplace/redhat-marketplace-p8w8l" Nov 29 05:25:55 crc kubenswrapper[4799]: I1129 05:25:55.664139 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcqmx\" (UniqueName: \"kubernetes.io/projected/c1e5797a-9ff4-466b-83c7-325510a31cf1-kube-api-access-mcqmx\") pod \"redhat-marketplace-p8w8l\" (UID: \"c1e5797a-9ff4-466b-83c7-325510a31cf1\") " pod="openshift-marketplace/redhat-marketplace-p8w8l" Nov 29 05:25:55 crc kubenswrapper[4799]: I1129 05:25:55.664217 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1e5797a-9ff4-466b-83c7-325510a31cf1-utilities\") pod \"redhat-marketplace-p8w8l\" (UID: \"c1e5797a-9ff4-466b-83c7-325510a31cf1\") " pod="openshift-marketplace/redhat-marketplace-p8w8l" Nov 29 05:25:55 crc kubenswrapper[4799]: I1129 05:25:55.766514 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1e5797a-9ff4-466b-83c7-325510a31cf1-catalog-content\") pod \"redhat-marketplace-p8w8l\" (UID: \"c1e5797a-9ff4-466b-83c7-325510a31cf1\") " pod="openshift-marketplace/redhat-marketplace-p8w8l" Nov 29 05:25:55 crc kubenswrapper[4799]: I1129 05:25:55.766634 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcqmx\" (UniqueName: \"kubernetes.io/projected/c1e5797a-9ff4-466b-83c7-325510a31cf1-kube-api-access-mcqmx\") pod \"redhat-marketplace-p8w8l\" (UID: \"c1e5797a-9ff4-466b-83c7-325510a31cf1\") " pod="openshift-marketplace/redhat-marketplace-p8w8l" Nov 29 05:25:55 crc kubenswrapper[4799]: I1129 05:25:55.766761 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1e5797a-9ff4-466b-83c7-325510a31cf1-utilities\") pod \"redhat-marketplace-p8w8l\" (UID: \"c1e5797a-9ff4-466b-83c7-325510a31cf1\") " pod="openshift-marketplace/redhat-marketplace-p8w8l" Nov 29 05:25:55 crc kubenswrapper[4799]: I1129 05:25:55.767260 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1e5797a-9ff4-466b-83c7-325510a31cf1-catalog-content\") pod \"redhat-marketplace-p8w8l\" (UID: \"c1e5797a-9ff4-466b-83c7-325510a31cf1\") " pod="openshift-marketplace/redhat-marketplace-p8w8l" Nov 29 05:25:55 crc kubenswrapper[4799]: I1129 05:25:55.768110 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1e5797a-9ff4-466b-83c7-325510a31cf1-utilities\") pod \"redhat-marketplace-p8w8l\" (UID: \"c1e5797a-9ff4-466b-83c7-325510a31cf1\") " pod="openshift-marketplace/redhat-marketplace-p8w8l" Nov 29 05:25:55 crc kubenswrapper[4799]: I1129 05:25:55.793375 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcqmx\" (UniqueName: \"kubernetes.io/projected/c1e5797a-9ff4-466b-83c7-325510a31cf1-kube-api-access-mcqmx\") pod \"redhat-marketplace-p8w8l\" (UID: \"c1e5797a-9ff4-466b-83c7-325510a31cf1\") " pod="openshift-marketplace/redhat-marketplace-p8w8l" Nov 29 05:25:55 crc kubenswrapper[4799]: I1129 05:25:55.862702 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p8w8l" Nov 29 05:25:56 crc kubenswrapper[4799]: I1129 05:25:56.402694 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-p8w8l"] Nov 29 05:25:56 crc kubenswrapper[4799]: I1129 05:25:56.786902 4799 generic.go:334] "Generic (PLEG): container finished" podID="c1e5797a-9ff4-466b-83c7-325510a31cf1" containerID="3693a060b97376f8dd38642836ae8583422fd52c1a295c0cc55fd38cda8a755c" exitCode=0 Nov 29 05:25:56 crc kubenswrapper[4799]: I1129 05:25:56.787082 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p8w8l" event={"ID":"c1e5797a-9ff4-466b-83c7-325510a31cf1","Type":"ContainerDied","Data":"3693a060b97376f8dd38642836ae8583422fd52c1a295c0cc55fd38cda8a755c"} Nov 29 05:25:56 crc kubenswrapper[4799]: I1129 05:25:56.788037 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p8w8l" event={"ID":"c1e5797a-9ff4-466b-83c7-325510a31cf1","Type":"ContainerStarted","Data":"047baba93ea79bb5918ee29a5ba4495bdd4fbdae92f5931d3774f3ff79fb2b89"} Nov 29 05:25:58 crc kubenswrapper[4799]: I1129 05:25:58.831629 4799 generic.go:334] "Generic (PLEG): container finished" podID="c1e5797a-9ff4-466b-83c7-325510a31cf1" containerID="22fe49fc803b16005d66ec08efc7242e6b2ca299e6e403580ca508d83ae29960" exitCode=0 Nov 29 05:25:58 crc kubenswrapper[4799]: I1129 05:25:58.831833 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p8w8l" event={"ID":"c1e5797a-9ff4-466b-83c7-325510a31cf1","Type":"ContainerDied","Data":"22fe49fc803b16005d66ec08efc7242e6b2ca299e6e403580ca508d83ae29960"} Nov 29 05:25:59 crc kubenswrapper[4799]: I1129 05:25:59.847241 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p8w8l" event={"ID":"c1e5797a-9ff4-466b-83c7-325510a31cf1","Type":"ContainerStarted","Data":"451d6aa536823416416a8e252cf17fead4e43941cf8c76a4c786e060f6a18724"} Nov 29 05:25:59 crc kubenswrapper[4799]: I1129 05:25:59.866547 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-p8w8l" podStartSLOduration=2.302762013 podStartE2EDuration="4.866521918s" podCreationTimestamp="2025-11-29 05:25:55 +0000 UTC" firstStartedPulling="2025-11-29 05:25:56.790451419 +0000 UTC m=+2832.433381819" lastFinishedPulling="2025-11-29 05:25:59.354211314 +0000 UTC m=+2834.997141724" observedRunningTime="2025-11-29 05:25:59.865464523 +0000 UTC m=+2835.508394953" watchObservedRunningTime="2025-11-29 05:25:59.866521918 +0000 UTC m=+2835.509452318" Nov 29 05:26:05 crc kubenswrapper[4799]: I1129 05:26:05.862910 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-p8w8l" Nov 29 05:26:05 crc kubenswrapper[4799]: I1129 05:26:05.864059 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-p8w8l" Nov 29 05:26:05 crc kubenswrapper[4799]: I1129 05:26:05.952286 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-p8w8l" Nov 29 05:26:06 crc kubenswrapper[4799]: I1129 05:26:06.047270 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-p8w8l" Nov 29 05:26:06 crc kubenswrapper[4799]: I1129 05:26:06.207764 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-p8w8l"] Nov 29 05:26:07 crc kubenswrapper[4799]: I1129 05:26:07.950280 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-p8w8l" podUID="c1e5797a-9ff4-466b-83c7-325510a31cf1" containerName="registry-server" containerID="cri-o://451d6aa536823416416a8e252cf17fead4e43941cf8c76a4c786e060f6a18724" gracePeriod=2 Nov 29 05:26:08 crc kubenswrapper[4799]: I1129 05:26:08.461086 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p8w8l" Nov 29 05:26:08 crc kubenswrapper[4799]: I1129 05:26:08.618142 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mcqmx\" (UniqueName: \"kubernetes.io/projected/c1e5797a-9ff4-466b-83c7-325510a31cf1-kube-api-access-mcqmx\") pod \"c1e5797a-9ff4-466b-83c7-325510a31cf1\" (UID: \"c1e5797a-9ff4-466b-83c7-325510a31cf1\") " Nov 29 05:26:08 crc kubenswrapper[4799]: I1129 05:26:08.618355 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1e5797a-9ff4-466b-83c7-325510a31cf1-utilities\") pod \"c1e5797a-9ff4-466b-83c7-325510a31cf1\" (UID: \"c1e5797a-9ff4-466b-83c7-325510a31cf1\") " Nov 29 05:26:08 crc kubenswrapper[4799]: I1129 05:26:08.618457 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1e5797a-9ff4-466b-83c7-325510a31cf1-catalog-content\") pod \"c1e5797a-9ff4-466b-83c7-325510a31cf1\" (UID: \"c1e5797a-9ff4-466b-83c7-325510a31cf1\") " Nov 29 05:26:08 crc kubenswrapper[4799]: I1129 05:26:08.621127 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1e5797a-9ff4-466b-83c7-325510a31cf1-utilities" (OuterVolumeSpecName: "utilities") pod "c1e5797a-9ff4-466b-83c7-325510a31cf1" (UID: "c1e5797a-9ff4-466b-83c7-325510a31cf1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:26:08 crc kubenswrapper[4799]: I1129 05:26:08.625965 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1e5797a-9ff4-466b-83c7-325510a31cf1-kube-api-access-mcqmx" (OuterVolumeSpecName: "kube-api-access-mcqmx") pod "c1e5797a-9ff4-466b-83c7-325510a31cf1" (UID: "c1e5797a-9ff4-466b-83c7-325510a31cf1"). InnerVolumeSpecName "kube-api-access-mcqmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:26:08 crc kubenswrapper[4799]: I1129 05:26:08.643319 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1e5797a-9ff4-466b-83c7-325510a31cf1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c1e5797a-9ff4-466b-83c7-325510a31cf1" (UID: "c1e5797a-9ff4-466b-83c7-325510a31cf1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:26:08 crc kubenswrapper[4799]: I1129 05:26:08.720998 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mcqmx\" (UniqueName: \"kubernetes.io/projected/c1e5797a-9ff4-466b-83c7-325510a31cf1-kube-api-access-mcqmx\") on node \"crc\" DevicePath \"\"" Nov 29 05:26:08 crc kubenswrapper[4799]: I1129 05:26:08.721030 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1e5797a-9ff4-466b-83c7-325510a31cf1-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 05:26:08 crc kubenswrapper[4799]: I1129 05:26:08.721041 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1e5797a-9ff4-466b-83c7-325510a31cf1-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 05:26:08 crc kubenswrapper[4799]: I1129 05:26:08.969433 4799 generic.go:334] "Generic (PLEG): container finished" podID="c1e5797a-9ff4-466b-83c7-325510a31cf1" containerID="451d6aa536823416416a8e252cf17fead4e43941cf8c76a4c786e060f6a18724" exitCode=0 Nov 29 05:26:08 crc kubenswrapper[4799]: I1129 05:26:08.969543 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p8w8l" event={"ID":"c1e5797a-9ff4-466b-83c7-325510a31cf1","Type":"ContainerDied","Data":"451d6aa536823416416a8e252cf17fead4e43941cf8c76a4c786e060f6a18724"} Nov 29 05:26:08 crc kubenswrapper[4799]: I1129 05:26:08.969628 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p8w8l" event={"ID":"c1e5797a-9ff4-466b-83c7-325510a31cf1","Type":"ContainerDied","Data":"047baba93ea79bb5918ee29a5ba4495bdd4fbdae92f5931d3774f3ff79fb2b89"} Nov 29 05:26:08 crc kubenswrapper[4799]: I1129 05:26:08.969631 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p8w8l" Nov 29 05:26:08 crc kubenswrapper[4799]: I1129 05:26:08.969660 4799 scope.go:117] "RemoveContainer" containerID="451d6aa536823416416a8e252cf17fead4e43941cf8c76a4c786e060f6a18724" Nov 29 05:26:09 crc kubenswrapper[4799]: I1129 05:26:09.003426 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-p8w8l"] Nov 29 05:26:09 crc kubenswrapper[4799]: I1129 05:26:09.014559 4799 scope.go:117] "RemoveContainer" containerID="22fe49fc803b16005d66ec08efc7242e6b2ca299e6e403580ca508d83ae29960" Nov 29 05:26:09 crc kubenswrapper[4799]: I1129 05:26:09.016906 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-p8w8l"] Nov 29 05:26:09 crc kubenswrapper[4799]: I1129 05:26:09.041212 4799 scope.go:117] "RemoveContainer" containerID="3693a060b97376f8dd38642836ae8583422fd52c1a295c0cc55fd38cda8a755c" Nov 29 05:26:09 crc kubenswrapper[4799]: I1129 05:26:09.080181 4799 scope.go:117] "RemoveContainer" containerID="451d6aa536823416416a8e252cf17fead4e43941cf8c76a4c786e060f6a18724" Nov 29 05:26:09 crc kubenswrapper[4799]: E1129 05:26:09.080864 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"451d6aa536823416416a8e252cf17fead4e43941cf8c76a4c786e060f6a18724\": container with ID starting with 451d6aa536823416416a8e252cf17fead4e43941cf8c76a4c786e060f6a18724 not found: ID does not exist" containerID="451d6aa536823416416a8e252cf17fead4e43941cf8c76a4c786e060f6a18724" Nov 29 05:26:09 crc kubenswrapper[4799]: I1129 05:26:09.080914 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"451d6aa536823416416a8e252cf17fead4e43941cf8c76a4c786e060f6a18724"} err="failed to get container status \"451d6aa536823416416a8e252cf17fead4e43941cf8c76a4c786e060f6a18724\": rpc error: code = NotFound desc = could not find container \"451d6aa536823416416a8e252cf17fead4e43941cf8c76a4c786e060f6a18724\": container with ID starting with 451d6aa536823416416a8e252cf17fead4e43941cf8c76a4c786e060f6a18724 not found: ID does not exist" Nov 29 05:26:09 crc kubenswrapper[4799]: I1129 05:26:09.080954 4799 scope.go:117] "RemoveContainer" containerID="22fe49fc803b16005d66ec08efc7242e6b2ca299e6e403580ca508d83ae29960" Nov 29 05:26:09 crc kubenswrapper[4799]: E1129 05:26:09.081525 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22fe49fc803b16005d66ec08efc7242e6b2ca299e6e403580ca508d83ae29960\": container with ID starting with 22fe49fc803b16005d66ec08efc7242e6b2ca299e6e403580ca508d83ae29960 not found: ID does not exist" containerID="22fe49fc803b16005d66ec08efc7242e6b2ca299e6e403580ca508d83ae29960" Nov 29 05:26:09 crc kubenswrapper[4799]: I1129 05:26:09.081565 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22fe49fc803b16005d66ec08efc7242e6b2ca299e6e403580ca508d83ae29960"} err="failed to get container status \"22fe49fc803b16005d66ec08efc7242e6b2ca299e6e403580ca508d83ae29960\": rpc error: code = NotFound desc = could not find container \"22fe49fc803b16005d66ec08efc7242e6b2ca299e6e403580ca508d83ae29960\": container with ID starting with 22fe49fc803b16005d66ec08efc7242e6b2ca299e6e403580ca508d83ae29960 not found: ID does not exist" Nov 29 05:26:09 crc kubenswrapper[4799]: I1129 05:26:09.081590 4799 scope.go:117] "RemoveContainer" containerID="3693a060b97376f8dd38642836ae8583422fd52c1a295c0cc55fd38cda8a755c" Nov 29 05:26:09 crc kubenswrapper[4799]: E1129 05:26:09.082099 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3693a060b97376f8dd38642836ae8583422fd52c1a295c0cc55fd38cda8a755c\": container with ID starting with 3693a060b97376f8dd38642836ae8583422fd52c1a295c0cc55fd38cda8a755c not found: ID does not exist" containerID="3693a060b97376f8dd38642836ae8583422fd52c1a295c0cc55fd38cda8a755c" Nov 29 05:26:09 crc kubenswrapper[4799]: I1129 05:26:09.082261 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3693a060b97376f8dd38642836ae8583422fd52c1a295c0cc55fd38cda8a755c"} err="failed to get container status \"3693a060b97376f8dd38642836ae8583422fd52c1a295c0cc55fd38cda8a755c\": rpc error: code = NotFound desc = could not find container \"3693a060b97376f8dd38642836ae8583422fd52c1a295c0cc55fd38cda8a755c\": container with ID starting with 3693a060b97376f8dd38642836ae8583422fd52c1a295c0cc55fd38cda8a755c not found: ID does not exist" Nov 29 05:26:10 crc kubenswrapper[4799]: I1129 05:26:10.676479 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1e5797a-9ff4-466b-83c7-325510a31cf1" path="/var/lib/kubelet/pods/c1e5797a-9ff4-466b-83c7-325510a31cf1/volumes" Nov 29 05:26:37 crc kubenswrapper[4799]: I1129 05:26:37.498227 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 05:26:37 crc kubenswrapper[4799]: I1129 05:26:37.499477 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 05:27:07 crc kubenswrapper[4799]: I1129 05:27:07.498452 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 05:27:07 crc kubenswrapper[4799]: I1129 05:27:07.499466 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 05:27:37 crc kubenswrapper[4799]: I1129 05:27:37.498427 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 05:27:37 crc kubenswrapper[4799]: I1129 05:27:37.499337 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 05:27:37 crc kubenswrapper[4799]: I1129 05:27:37.499395 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-26t88" Nov 29 05:27:37 crc kubenswrapper[4799]: I1129 05:27:37.500350 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e6d4b3d1d5b5e1eeb13508d67de3d3c3aac52e76337b55e53793185c0d001d82"} pod="openshift-machine-config-operator/machine-config-daemon-26t88" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 29 05:27:37 crc kubenswrapper[4799]: I1129 05:27:37.500404 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" containerID="cri-o://e6d4b3d1d5b5e1eeb13508d67de3d3c3aac52e76337b55e53793185c0d001d82" gracePeriod=600 Nov 29 05:27:37 crc kubenswrapper[4799]: E1129 05:27:37.639103 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:27:38 crc kubenswrapper[4799]: I1129 05:27:38.120572 4799 generic.go:334] "Generic (PLEG): container finished" podID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerID="e6d4b3d1d5b5e1eeb13508d67de3d3c3aac52e76337b55e53793185c0d001d82" exitCode=0 Nov 29 05:27:38 crc kubenswrapper[4799]: I1129 05:27:38.120689 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" event={"ID":"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8","Type":"ContainerDied","Data":"e6d4b3d1d5b5e1eeb13508d67de3d3c3aac52e76337b55e53793185c0d001d82"} Nov 29 05:27:38 crc kubenswrapper[4799]: I1129 05:27:38.121239 4799 scope.go:117] "RemoveContainer" containerID="4a42289995c12e4c41519c0c91fe9e1ab67dd9e0388b75fa426811cd1920c88e" Nov 29 05:27:38 crc kubenswrapper[4799]: I1129 05:27:38.122311 4799 scope.go:117] "RemoveContainer" containerID="e6d4b3d1d5b5e1eeb13508d67de3d3c3aac52e76337b55e53793185c0d001d82" Nov 29 05:27:38 crc kubenswrapper[4799]: E1129 05:27:38.122994 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:27:53 crc kubenswrapper[4799]: I1129 05:27:53.660558 4799 scope.go:117] "RemoveContainer" containerID="e6d4b3d1d5b5e1eeb13508d67de3d3c3aac52e76337b55e53793185c0d001d82" Nov 29 05:27:53 crc kubenswrapper[4799]: E1129 05:27:53.661993 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:28:05 crc kubenswrapper[4799]: I1129 05:28:05.659451 4799 scope.go:117] "RemoveContainer" containerID="e6d4b3d1d5b5e1eeb13508d67de3d3c3aac52e76337b55e53793185c0d001d82" Nov 29 05:28:05 crc kubenswrapper[4799]: E1129 05:28:05.661001 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:28:17 crc kubenswrapper[4799]: I1129 05:28:17.659856 4799 scope.go:117] "RemoveContainer" containerID="e6d4b3d1d5b5e1eeb13508d67de3d3c3aac52e76337b55e53793185c0d001d82" Nov 29 05:28:17 crc kubenswrapper[4799]: E1129 05:28:17.661200 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:28:25 crc kubenswrapper[4799]: I1129 05:28:25.695343 4799 generic.go:334] "Generic (PLEG): container finished" podID="26a0b212-c652-4475-8210-4b62b653ed79" containerID="94d4a5e46c1fefdf618e24b6387dae4a821483a636fec1ff9ea7faeb4dbf2b69" exitCode=0 Nov 29 05:28:25 crc kubenswrapper[4799]: I1129 05:28:25.695392 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b" event={"ID":"26a0b212-c652-4475-8210-4b62b653ed79","Type":"ContainerDied","Data":"94d4a5e46c1fefdf618e24b6387dae4a821483a636fec1ff9ea7faeb4dbf2b69"} Nov 29 05:28:27 crc kubenswrapper[4799]: I1129 05:28:27.307299 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b" Nov 29 05:28:27 crc kubenswrapper[4799]: I1129 05:28:27.464133 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ttvf\" (UniqueName: \"kubernetes.io/projected/26a0b212-c652-4475-8210-4b62b653ed79-kube-api-access-7ttvf\") pod \"26a0b212-c652-4475-8210-4b62b653ed79\" (UID: \"26a0b212-c652-4475-8210-4b62b653ed79\") " Nov 29 05:28:27 crc kubenswrapper[4799]: I1129 05:28:27.464199 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/26a0b212-c652-4475-8210-4b62b653ed79-nova-extra-config-0\") pod \"26a0b212-c652-4475-8210-4b62b653ed79\" (UID: \"26a0b212-c652-4475-8210-4b62b653ed79\") " Nov 29 05:28:27 crc kubenswrapper[4799]: I1129 05:28:27.464233 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26a0b212-c652-4475-8210-4b62b653ed79-nova-custom-ceph-combined-ca-bundle\") pod \"26a0b212-c652-4475-8210-4b62b653ed79\" (UID: \"26a0b212-c652-4475-8210-4b62b653ed79\") " Nov 29 05:28:27 crc kubenswrapper[4799]: I1129 05:28:27.464298 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/26a0b212-c652-4475-8210-4b62b653ed79-nova-cell1-compute-config-1\") pod \"26a0b212-c652-4475-8210-4b62b653ed79\" (UID: \"26a0b212-c652-4475-8210-4b62b653ed79\") " Nov 29 05:28:27 crc kubenswrapper[4799]: I1129 05:28:27.464323 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/26a0b212-c652-4475-8210-4b62b653ed79-inventory\") pod \"26a0b212-c652-4475-8210-4b62b653ed79\" (UID: \"26a0b212-c652-4475-8210-4b62b653ed79\") " Nov 29 05:28:27 crc kubenswrapper[4799]: I1129 05:28:27.464458 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/26a0b212-c652-4475-8210-4b62b653ed79-nova-migration-ssh-key-0\") pod \"26a0b212-c652-4475-8210-4b62b653ed79\" (UID: \"26a0b212-c652-4475-8210-4b62b653ed79\") " Nov 29 05:28:27 crc kubenswrapper[4799]: I1129 05:28:27.464532 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/26a0b212-c652-4475-8210-4b62b653ed79-ceph\") pod \"26a0b212-c652-4475-8210-4b62b653ed79\" (UID: \"26a0b212-c652-4475-8210-4b62b653ed79\") " Nov 29 05:28:27 crc kubenswrapper[4799]: I1129 05:28:27.465225 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/26a0b212-c652-4475-8210-4b62b653ed79-ceph-nova-0\") pod \"26a0b212-c652-4475-8210-4b62b653ed79\" (UID: \"26a0b212-c652-4475-8210-4b62b653ed79\") " Nov 29 05:28:27 crc kubenswrapper[4799]: I1129 05:28:27.465320 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/26a0b212-c652-4475-8210-4b62b653ed79-ssh-key\") pod \"26a0b212-c652-4475-8210-4b62b653ed79\" (UID: \"26a0b212-c652-4475-8210-4b62b653ed79\") " Nov 29 05:28:27 crc kubenswrapper[4799]: I1129 05:28:27.465481 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/26a0b212-c652-4475-8210-4b62b653ed79-nova-cell1-compute-config-0\") pod \"26a0b212-c652-4475-8210-4b62b653ed79\" (UID: \"26a0b212-c652-4475-8210-4b62b653ed79\") " Nov 29 05:28:27 crc kubenswrapper[4799]: I1129 05:28:27.465869 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/26a0b212-c652-4475-8210-4b62b653ed79-nova-migration-ssh-key-1\") pod \"26a0b212-c652-4475-8210-4b62b653ed79\" (UID: \"26a0b212-c652-4475-8210-4b62b653ed79\") " Nov 29 05:28:27 crc kubenswrapper[4799]: I1129 05:28:27.472325 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26a0b212-c652-4475-8210-4b62b653ed79-nova-custom-ceph-combined-ca-bundle" (OuterVolumeSpecName: "nova-custom-ceph-combined-ca-bundle") pod "26a0b212-c652-4475-8210-4b62b653ed79" (UID: "26a0b212-c652-4475-8210-4b62b653ed79"). InnerVolumeSpecName "nova-custom-ceph-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:28:27 crc kubenswrapper[4799]: I1129 05:28:27.473958 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26a0b212-c652-4475-8210-4b62b653ed79-ceph" (OuterVolumeSpecName: "ceph") pod "26a0b212-c652-4475-8210-4b62b653ed79" (UID: "26a0b212-c652-4475-8210-4b62b653ed79"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:28:27 crc kubenswrapper[4799]: I1129 05:28:27.475115 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26a0b212-c652-4475-8210-4b62b653ed79-kube-api-access-7ttvf" (OuterVolumeSpecName: "kube-api-access-7ttvf") pod "26a0b212-c652-4475-8210-4b62b653ed79" (UID: "26a0b212-c652-4475-8210-4b62b653ed79"). InnerVolumeSpecName "kube-api-access-7ttvf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:28:27 crc kubenswrapper[4799]: I1129 05:28:27.499222 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26a0b212-c652-4475-8210-4b62b653ed79-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "26a0b212-c652-4475-8210-4b62b653ed79" (UID: "26a0b212-c652-4475-8210-4b62b653ed79"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:28:27 crc kubenswrapper[4799]: I1129 05:28:27.501635 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26a0b212-c652-4475-8210-4b62b653ed79-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "26a0b212-c652-4475-8210-4b62b653ed79" (UID: "26a0b212-c652-4475-8210-4b62b653ed79"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:28:27 crc kubenswrapper[4799]: I1129 05:28:27.502365 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26a0b212-c652-4475-8210-4b62b653ed79-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "26a0b212-c652-4475-8210-4b62b653ed79" (UID: "26a0b212-c652-4475-8210-4b62b653ed79"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:28:27 crc kubenswrapper[4799]: I1129 05:28:27.504335 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26a0b212-c652-4475-8210-4b62b653ed79-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "26a0b212-c652-4475-8210-4b62b653ed79" (UID: "26a0b212-c652-4475-8210-4b62b653ed79"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:28:27 crc kubenswrapper[4799]: I1129 05:28:27.512457 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26a0b212-c652-4475-8210-4b62b653ed79-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "26a0b212-c652-4475-8210-4b62b653ed79" (UID: "26a0b212-c652-4475-8210-4b62b653ed79"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 05:28:27 crc kubenswrapper[4799]: I1129 05:28:27.517374 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26a0b212-c652-4475-8210-4b62b653ed79-inventory" (OuterVolumeSpecName: "inventory") pod "26a0b212-c652-4475-8210-4b62b653ed79" (UID: "26a0b212-c652-4475-8210-4b62b653ed79"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:28:27 crc kubenswrapper[4799]: I1129 05:28:27.527512 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26a0b212-c652-4475-8210-4b62b653ed79-ceph-nova-0" (OuterVolumeSpecName: "ceph-nova-0") pod "26a0b212-c652-4475-8210-4b62b653ed79" (UID: "26a0b212-c652-4475-8210-4b62b653ed79"). InnerVolumeSpecName "ceph-nova-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 05:28:27 crc kubenswrapper[4799]: I1129 05:28:27.532735 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26a0b212-c652-4475-8210-4b62b653ed79-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "26a0b212-c652-4475-8210-4b62b653ed79" (UID: "26a0b212-c652-4475-8210-4b62b653ed79"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:28:27 crc kubenswrapper[4799]: I1129 05:28:27.569840 4799 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/26a0b212-c652-4475-8210-4b62b653ed79-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Nov 29 05:28:27 crc kubenswrapper[4799]: I1129 05:28:27.569909 4799 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/26a0b212-c652-4475-8210-4b62b653ed79-ceph\") on node \"crc\" DevicePath \"\"" Nov 29 05:28:27 crc kubenswrapper[4799]: I1129 05:28:27.569933 4799 reconciler_common.go:293] "Volume detached for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/26a0b212-c652-4475-8210-4b62b653ed79-ceph-nova-0\") on node \"crc\" DevicePath \"\"" Nov 29 05:28:27 crc kubenswrapper[4799]: I1129 05:28:27.569956 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/26a0b212-c652-4475-8210-4b62b653ed79-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 29 05:28:27 crc kubenswrapper[4799]: I1129 05:28:27.569974 4799 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/26a0b212-c652-4475-8210-4b62b653ed79-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Nov 29 05:28:27 crc kubenswrapper[4799]: I1129 05:28:27.569995 4799 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/26a0b212-c652-4475-8210-4b62b653ed79-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Nov 29 05:28:27 crc kubenswrapper[4799]: I1129 05:28:27.570013 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ttvf\" (UniqueName: \"kubernetes.io/projected/26a0b212-c652-4475-8210-4b62b653ed79-kube-api-access-7ttvf\") on node \"crc\" DevicePath \"\"" Nov 29 05:28:27 crc kubenswrapper[4799]: I1129 05:28:27.570031 4799 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/26a0b212-c652-4475-8210-4b62b653ed79-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Nov 29 05:28:27 crc kubenswrapper[4799]: I1129 05:28:27.570050 4799 reconciler_common.go:293] "Volume detached for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26a0b212-c652-4475-8210-4b62b653ed79-nova-custom-ceph-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 05:28:27 crc kubenswrapper[4799]: I1129 05:28:27.570081 4799 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/26a0b212-c652-4475-8210-4b62b653ed79-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Nov 29 05:28:27 crc kubenswrapper[4799]: I1129 05:28:27.570106 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/26a0b212-c652-4475-8210-4b62b653ed79-inventory\") on node \"crc\" DevicePath \"\"" Nov 29 05:28:27 crc kubenswrapper[4799]: I1129 05:28:27.720916 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b" event={"ID":"26a0b212-c652-4475-8210-4b62b653ed79","Type":"ContainerDied","Data":"afe9ac99e8c014634b44f791e0d91e88a2f77af7e01c558b56355a9116f38e4b"} Nov 29 05:28:27 crc kubenswrapper[4799]: I1129 05:28:27.721006 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b" Nov 29 05:28:27 crc kubenswrapper[4799]: I1129 05:28:27.721036 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="afe9ac99e8c014634b44f791e0d91e88a2f77af7e01c558b56355a9116f38e4b" Nov 29 05:28:29 crc kubenswrapper[4799]: I1129 05:28:29.660720 4799 scope.go:117] "RemoveContainer" containerID="e6d4b3d1d5b5e1eeb13508d67de3d3c3aac52e76337b55e53793185c0d001d82" Nov 29 05:28:29 crc kubenswrapper[4799]: E1129 05:28:29.662132 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.571077 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-volume1-0"] Nov 29 05:28:42 crc kubenswrapper[4799]: E1129 05:28:42.574481 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26a0b212-c652-4475-8210-4b62b653ed79" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.574521 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="26a0b212-c652-4475-8210-4b62b653ed79" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Nov 29 05:28:42 crc kubenswrapper[4799]: E1129 05:28:42.574538 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1e5797a-9ff4-466b-83c7-325510a31cf1" containerName="extract-utilities" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.574548 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1e5797a-9ff4-466b-83c7-325510a31cf1" containerName="extract-utilities" Nov 29 05:28:42 crc kubenswrapper[4799]: E1129 05:28:42.574568 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1e5797a-9ff4-466b-83c7-325510a31cf1" containerName="registry-server" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.574577 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1e5797a-9ff4-466b-83c7-325510a31cf1" containerName="registry-server" Nov 29 05:28:42 crc kubenswrapper[4799]: E1129 05:28:42.574617 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1e5797a-9ff4-466b-83c7-325510a31cf1" containerName="extract-content" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.574623 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1e5797a-9ff4-466b-83c7-325510a31cf1" containerName="extract-content" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.575088 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="26a0b212-c652-4475-8210-4b62b653ed79" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.575110 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1e5797a-9ff4-466b-83c7-325510a31cf1" containerName="registry-server" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.576477 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.578566 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-volume1-config-data" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.578771 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.600126 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.609393 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.611537 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.619757 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.625190 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.640027 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0907967b-0bf5-4fe0-9862-ffbd54efde7f-run\") pod \"cinder-volume-volume1-0\" (UID: \"0907967b-0bf5-4fe0-9862-ffbd54efde7f\") " pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.640139 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0907967b-0bf5-4fe0-9862-ffbd54efde7f-dev\") pod \"cinder-volume-volume1-0\" (UID: \"0907967b-0bf5-4fe0-9862-ffbd54efde7f\") " pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.640221 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0907967b-0bf5-4fe0-9862-ffbd54efde7f-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"0907967b-0bf5-4fe0-9862-ffbd54efde7f\") " pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.640280 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0907967b-0bf5-4fe0-9862-ffbd54efde7f-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"0907967b-0bf5-4fe0-9862-ffbd54efde7f\") " pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.640367 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0907967b-0bf5-4fe0-9862-ffbd54efde7f-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"0907967b-0bf5-4fe0-9862-ffbd54efde7f\") " pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.640402 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0907967b-0bf5-4fe0-9862-ffbd54efde7f-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"0907967b-0bf5-4fe0-9862-ffbd54efde7f\") " pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.640509 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0907967b-0bf5-4fe0-9862-ffbd54efde7f-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"0907967b-0bf5-4fe0-9862-ffbd54efde7f\") " pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.640597 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pr5s5\" (UniqueName: \"kubernetes.io/projected/0907967b-0bf5-4fe0-9862-ffbd54efde7f-kube-api-access-pr5s5\") pod \"cinder-volume-volume1-0\" (UID: \"0907967b-0bf5-4fe0-9862-ffbd54efde7f\") " pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.640649 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0907967b-0bf5-4fe0-9862-ffbd54efde7f-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"0907967b-0bf5-4fe0-9862-ffbd54efde7f\") " pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.640674 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0907967b-0bf5-4fe0-9862-ffbd54efde7f-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"0907967b-0bf5-4fe0-9862-ffbd54efde7f\") " pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.640749 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0907967b-0bf5-4fe0-9862-ffbd54efde7f-sys\") pod \"cinder-volume-volume1-0\" (UID: \"0907967b-0bf5-4fe0-9862-ffbd54efde7f\") " pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.640771 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/0907967b-0bf5-4fe0-9862-ffbd54efde7f-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"0907967b-0bf5-4fe0-9862-ffbd54efde7f\") " pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.641933 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0907967b-0bf5-4fe0-9862-ffbd54efde7f-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"0907967b-0bf5-4fe0-9862-ffbd54efde7f\") " pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.642023 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0907967b-0bf5-4fe0-9862-ffbd54efde7f-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"0907967b-0bf5-4fe0-9862-ffbd54efde7f\") " pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.642128 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/0907967b-0bf5-4fe0-9862-ffbd54efde7f-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"0907967b-0bf5-4fe0-9862-ffbd54efde7f\") " pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.642154 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0907967b-0bf5-4fe0-9862-ffbd54efde7f-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"0907967b-0bf5-4fe0-9862-ffbd54efde7f\") " pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.744156 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0907967b-0bf5-4fe0-9862-ffbd54efde7f-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"0907967b-0bf5-4fe0-9862-ffbd54efde7f\") " pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.744580 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0907967b-0bf5-4fe0-9862-ffbd54efde7f-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"0907967b-0bf5-4fe0-9862-ffbd54efde7f\") " pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.744290 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0907967b-0bf5-4fe0-9862-ffbd54efde7f-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"0907967b-0bf5-4fe0-9862-ffbd54efde7f\") " pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.744617 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/caa349ef-0de7-46c4-a7cd-fdf43aa6f949-scripts\") pod \"cinder-backup-0\" (UID: \"caa349ef-0de7-46c4-a7cd-fdf43aa6f949\") " pod="openstack/cinder-backup-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.744682 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0907967b-0bf5-4fe0-9862-ffbd54efde7f-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"0907967b-0bf5-4fe0-9862-ffbd54efde7f\") " pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.744743 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pr5s5\" (UniqueName: \"kubernetes.io/projected/0907967b-0bf5-4fe0-9862-ffbd54efde7f-kube-api-access-pr5s5\") pod \"cinder-volume-volume1-0\" (UID: \"0907967b-0bf5-4fe0-9862-ffbd54efde7f\") " pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.744766 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/caa349ef-0de7-46c4-a7cd-fdf43aa6f949-dev\") pod \"cinder-backup-0\" (UID: \"caa349ef-0de7-46c4-a7cd-fdf43aa6f949\") " pod="openstack/cinder-backup-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.744806 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/caa349ef-0de7-46c4-a7cd-fdf43aa6f949-config-data-custom\") pod \"cinder-backup-0\" (UID: \"caa349ef-0de7-46c4-a7cd-fdf43aa6f949\") " pod="openstack/cinder-backup-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.744836 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0907967b-0bf5-4fe0-9862-ffbd54efde7f-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"0907967b-0bf5-4fe0-9862-ffbd54efde7f\") " pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.744857 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0907967b-0bf5-4fe0-9862-ffbd54efde7f-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"0907967b-0bf5-4fe0-9862-ffbd54efde7f\") " pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.744876 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/caa349ef-0de7-46c4-a7cd-fdf43aa6f949-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"caa349ef-0de7-46c4-a7cd-fdf43aa6f949\") " pod="openstack/cinder-backup-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.744907 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/caa349ef-0de7-46c4-a7cd-fdf43aa6f949-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"caa349ef-0de7-46c4-a7cd-fdf43aa6f949\") " pod="openstack/cinder-backup-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.744936 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7d74t\" (UniqueName: \"kubernetes.io/projected/caa349ef-0de7-46c4-a7cd-fdf43aa6f949-kube-api-access-7d74t\") pod \"cinder-backup-0\" (UID: \"caa349ef-0de7-46c4-a7cd-fdf43aa6f949\") " pod="openstack/cinder-backup-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.744994 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0907967b-0bf5-4fe0-9862-ffbd54efde7f-sys\") pod \"cinder-volume-volume1-0\" (UID: \"0907967b-0bf5-4fe0-9862-ffbd54efde7f\") " pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.745016 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/0907967b-0bf5-4fe0-9862-ffbd54efde7f-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"0907967b-0bf5-4fe0-9862-ffbd54efde7f\") " pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.745067 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/caa349ef-0de7-46c4-a7cd-fdf43aa6f949-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"caa349ef-0de7-46c4-a7cd-fdf43aa6f949\") " pod="openstack/cinder-backup-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.745130 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0907967b-0bf5-4fe0-9862-ffbd54efde7f-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"0907967b-0bf5-4fe0-9862-ffbd54efde7f\") " pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.745152 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0907967b-0bf5-4fe0-9862-ffbd54efde7f-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"0907967b-0bf5-4fe0-9862-ffbd54efde7f\") " pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.745173 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/caa349ef-0de7-46c4-a7cd-fdf43aa6f949-run\") pod \"cinder-backup-0\" (UID: \"caa349ef-0de7-46c4-a7cd-fdf43aa6f949\") " pod="openstack/cinder-backup-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.745202 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/caa349ef-0de7-46c4-a7cd-fdf43aa6f949-config-data\") pod \"cinder-backup-0\" (UID: \"caa349ef-0de7-46c4-a7cd-fdf43aa6f949\") " pod="openstack/cinder-backup-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.745224 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/caa349ef-0de7-46c4-a7cd-fdf43aa6f949-sys\") pod \"cinder-backup-0\" (UID: \"caa349ef-0de7-46c4-a7cd-fdf43aa6f949\") " pod="openstack/cinder-backup-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.745251 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/0907967b-0bf5-4fe0-9862-ffbd54efde7f-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"0907967b-0bf5-4fe0-9862-ffbd54efde7f\") " pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.745255 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0907967b-0bf5-4fe0-9862-ffbd54efde7f-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"0907967b-0bf5-4fe0-9862-ffbd54efde7f\") " pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.745271 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0907967b-0bf5-4fe0-9862-ffbd54efde7f-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"0907967b-0bf5-4fe0-9862-ffbd54efde7f\") " pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.745345 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0907967b-0bf5-4fe0-9862-ffbd54efde7f-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"0907967b-0bf5-4fe0-9862-ffbd54efde7f\") " pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.745465 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0907967b-0bf5-4fe0-9862-ffbd54efde7f-sys\") pod \"cinder-volume-volume1-0\" (UID: \"0907967b-0bf5-4fe0-9862-ffbd54efde7f\") " pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.745633 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/0907967b-0bf5-4fe0-9862-ffbd54efde7f-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"0907967b-0bf5-4fe0-9862-ffbd54efde7f\") " pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.745663 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0907967b-0bf5-4fe0-9862-ffbd54efde7f-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"0907967b-0bf5-4fe0-9862-ffbd54efde7f\") " pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.745753 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/0907967b-0bf5-4fe0-9862-ffbd54efde7f-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"0907967b-0bf5-4fe0-9862-ffbd54efde7f\") " pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.746047 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/caa349ef-0de7-46c4-a7cd-fdf43aa6f949-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"caa349ef-0de7-46c4-a7cd-fdf43aa6f949\") " pod="openstack/cinder-backup-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.746094 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/caa349ef-0de7-46c4-a7cd-fdf43aa6f949-lib-modules\") pod \"cinder-backup-0\" (UID: \"caa349ef-0de7-46c4-a7cd-fdf43aa6f949\") " pod="openstack/cinder-backup-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.746208 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0907967b-0bf5-4fe0-9862-ffbd54efde7f-run\") pod \"cinder-volume-volume1-0\" (UID: \"0907967b-0bf5-4fe0-9862-ffbd54efde7f\") " pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.746255 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/caa349ef-0de7-46c4-a7cd-fdf43aa6f949-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"caa349ef-0de7-46c4-a7cd-fdf43aa6f949\") " pod="openstack/cinder-backup-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.746294 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/caa349ef-0de7-46c4-a7cd-fdf43aa6f949-etc-nvme\") pod \"cinder-backup-0\" (UID: \"caa349ef-0de7-46c4-a7cd-fdf43aa6f949\") " pod="openstack/cinder-backup-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.746328 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0907967b-0bf5-4fe0-9862-ffbd54efde7f-dev\") pod \"cinder-volume-volume1-0\" (UID: \"0907967b-0bf5-4fe0-9862-ffbd54efde7f\") " pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.746446 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0907967b-0bf5-4fe0-9862-ffbd54efde7f-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"0907967b-0bf5-4fe0-9862-ffbd54efde7f\") " pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.746472 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/caa349ef-0de7-46c4-a7cd-fdf43aa6f949-ceph\") pod \"cinder-backup-0\" (UID: \"caa349ef-0de7-46c4-a7cd-fdf43aa6f949\") " pod="openstack/cinder-backup-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.746529 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0907967b-0bf5-4fe0-9862-ffbd54efde7f-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"0907967b-0bf5-4fe0-9862-ffbd54efde7f\") " pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.746567 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0907967b-0bf5-4fe0-9862-ffbd54efde7f-dev\") pod \"cinder-volume-volume1-0\" (UID: \"0907967b-0bf5-4fe0-9862-ffbd54efde7f\") " pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.746615 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0907967b-0bf5-4fe0-9862-ffbd54efde7f-run\") pod \"cinder-volume-volume1-0\" (UID: \"0907967b-0bf5-4fe0-9862-ffbd54efde7f\") " pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.746627 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/caa349ef-0de7-46c4-a7cd-fdf43aa6f949-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"caa349ef-0de7-46c4-a7cd-fdf43aa6f949\") " pod="openstack/cinder-backup-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.746765 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0907967b-0bf5-4fe0-9862-ffbd54efde7f-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"0907967b-0bf5-4fe0-9862-ffbd54efde7f\") " pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.756535 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0907967b-0bf5-4fe0-9862-ffbd54efde7f-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"0907967b-0bf5-4fe0-9862-ffbd54efde7f\") " pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.756657 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0907967b-0bf5-4fe0-9862-ffbd54efde7f-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"0907967b-0bf5-4fe0-9862-ffbd54efde7f\") " pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.757263 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0907967b-0bf5-4fe0-9862-ffbd54efde7f-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"0907967b-0bf5-4fe0-9862-ffbd54efde7f\") " pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.763440 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0907967b-0bf5-4fe0-9862-ffbd54efde7f-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"0907967b-0bf5-4fe0-9862-ffbd54efde7f\") " pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.779302 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0907967b-0bf5-4fe0-9862-ffbd54efde7f-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"0907967b-0bf5-4fe0-9862-ffbd54efde7f\") " pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.785757 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pr5s5\" (UniqueName: \"kubernetes.io/projected/0907967b-0bf5-4fe0-9862-ffbd54efde7f-kube-api-access-pr5s5\") pod \"cinder-volume-volume1-0\" (UID: \"0907967b-0bf5-4fe0-9862-ffbd54efde7f\") " pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.849502 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/caa349ef-0de7-46c4-a7cd-fdf43aa6f949-etc-nvme\") pod \"cinder-backup-0\" (UID: \"caa349ef-0de7-46c4-a7cd-fdf43aa6f949\") " pod="openstack/cinder-backup-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.849593 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/caa349ef-0de7-46c4-a7cd-fdf43aa6f949-ceph\") pod \"cinder-backup-0\" (UID: \"caa349ef-0de7-46c4-a7cd-fdf43aa6f949\") " pod="openstack/cinder-backup-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.849629 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/caa349ef-0de7-46c4-a7cd-fdf43aa6f949-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"caa349ef-0de7-46c4-a7cd-fdf43aa6f949\") " pod="openstack/cinder-backup-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.849681 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/caa349ef-0de7-46c4-a7cd-fdf43aa6f949-scripts\") pod \"cinder-backup-0\" (UID: \"caa349ef-0de7-46c4-a7cd-fdf43aa6f949\") " pod="openstack/cinder-backup-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.849715 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/caa349ef-0de7-46c4-a7cd-fdf43aa6f949-dev\") pod \"cinder-backup-0\" (UID: \"caa349ef-0de7-46c4-a7cd-fdf43aa6f949\") " pod="openstack/cinder-backup-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.849732 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/caa349ef-0de7-46c4-a7cd-fdf43aa6f949-config-data-custom\") pod \"cinder-backup-0\" (UID: \"caa349ef-0de7-46c4-a7cd-fdf43aa6f949\") " pod="openstack/cinder-backup-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.849750 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/caa349ef-0de7-46c4-a7cd-fdf43aa6f949-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"caa349ef-0de7-46c4-a7cd-fdf43aa6f949\") " pod="openstack/cinder-backup-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.849773 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/caa349ef-0de7-46c4-a7cd-fdf43aa6f949-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"caa349ef-0de7-46c4-a7cd-fdf43aa6f949\") " pod="openstack/cinder-backup-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.849805 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7d74t\" (UniqueName: \"kubernetes.io/projected/caa349ef-0de7-46c4-a7cd-fdf43aa6f949-kube-api-access-7d74t\") pod \"cinder-backup-0\" (UID: \"caa349ef-0de7-46c4-a7cd-fdf43aa6f949\") " pod="openstack/cinder-backup-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.849842 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/caa349ef-0de7-46c4-a7cd-fdf43aa6f949-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"caa349ef-0de7-46c4-a7cd-fdf43aa6f949\") " pod="openstack/cinder-backup-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.849869 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/caa349ef-0de7-46c4-a7cd-fdf43aa6f949-run\") pod \"cinder-backup-0\" (UID: \"caa349ef-0de7-46c4-a7cd-fdf43aa6f949\") " pod="openstack/cinder-backup-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.849892 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/caa349ef-0de7-46c4-a7cd-fdf43aa6f949-config-data\") pod \"cinder-backup-0\" (UID: \"caa349ef-0de7-46c4-a7cd-fdf43aa6f949\") " pod="openstack/cinder-backup-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.849912 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/caa349ef-0de7-46c4-a7cd-fdf43aa6f949-sys\") pod \"cinder-backup-0\" (UID: \"caa349ef-0de7-46c4-a7cd-fdf43aa6f949\") " pod="openstack/cinder-backup-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.849945 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/caa349ef-0de7-46c4-a7cd-fdf43aa6f949-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"caa349ef-0de7-46c4-a7cd-fdf43aa6f949\") " pod="openstack/cinder-backup-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.849962 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/caa349ef-0de7-46c4-a7cd-fdf43aa6f949-lib-modules\") pod \"cinder-backup-0\" (UID: \"caa349ef-0de7-46c4-a7cd-fdf43aa6f949\") " pod="openstack/cinder-backup-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.849986 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/caa349ef-0de7-46c4-a7cd-fdf43aa6f949-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"caa349ef-0de7-46c4-a7cd-fdf43aa6f949\") " pod="openstack/cinder-backup-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.850065 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/caa349ef-0de7-46c4-a7cd-fdf43aa6f949-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"caa349ef-0de7-46c4-a7cd-fdf43aa6f949\") " pod="openstack/cinder-backup-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.850205 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/caa349ef-0de7-46c4-a7cd-fdf43aa6f949-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"caa349ef-0de7-46c4-a7cd-fdf43aa6f949\") " pod="openstack/cinder-backup-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.850298 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/caa349ef-0de7-46c4-a7cd-fdf43aa6f949-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"caa349ef-0de7-46c4-a7cd-fdf43aa6f949\") " pod="openstack/cinder-backup-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.850438 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/caa349ef-0de7-46c4-a7cd-fdf43aa6f949-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"caa349ef-0de7-46c4-a7cd-fdf43aa6f949\") " pod="openstack/cinder-backup-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.850470 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/caa349ef-0de7-46c4-a7cd-fdf43aa6f949-run\") pod \"cinder-backup-0\" (UID: \"caa349ef-0de7-46c4-a7cd-fdf43aa6f949\") " pod="openstack/cinder-backup-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.850694 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/caa349ef-0de7-46c4-a7cd-fdf43aa6f949-sys\") pod \"cinder-backup-0\" (UID: \"caa349ef-0de7-46c4-a7cd-fdf43aa6f949\") " pod="openstack/cinder-backup-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.850746 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/caa349ef-0de7-46c4-a7cd-fdf43aa6f949-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"caa349ef-0de7-46c4-a7cd-fdf43aa6f949\") " pod="openstack/cinder-backup-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.850780 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/caa349ef-0de7-46c4-a7cd-fdf43aa6f949-lib-modules\") pod \"cinder-backup-0\" (UID: \"caa349ef-0de7-46c4-a7cd-fdf43aa6f949\") " pod="openstack/cinder-backup-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.850839 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/caa349ef-0de7-46c4-a7cd-fdf43aa6f949-etc-nvme\") pod \"cinder-backup-0\" (UID: \"caa349ef-0de7-46c4-a7cd-fdf43aa6f949\") " pod="openstack/cinder-backup-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.850863 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/caa349ef-0de7-46c4-a7cd-fdf43aa6f949-dev\") pod \"cinder-backup-0\" (UID: \"caa349ef-0de7-46c4-a7cd-fdf43aa6f949\") " pod="openstack/cinder-backup-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.854835 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/caa349ef-0de7-46c4-a7cd-fdf43aa6f949-scripts\") pod \"cinder-backup-0\" (UID: \"caa349ef-0de7-46c4-a7cd-fdf43aa6f949\") " pod="openstack/cinder-backup-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.855506 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/caa349ef-0de7-46c4-a7cd-fdf43aa6f949-config-data\") pod \"cinder-backup-0\" (UID: \"caa349ef-0de7-46c4-a7cd-fdf43aa6f949\") " pod="openstack/cinder-backup-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.856516 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/caa349ef-0de7-46c4-a7cd-fdf43aa6f949-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"caa349ef-0de7-46c4-a7cd-fdf43aa6f949\") " pod="openstack/cinder-backup-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.862412 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/caa349ef-0de7-46c4-a7cd-fdf43aa6f949-config-data-custom\") pod \"cinder-backup-0\" (UID: \"caa349ef-0de7-46c4-a7cd-fdf43aa6f949\") " pod="openstack/cinder-backup-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.874027 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/caa349ef-0de7-46c4-a7cd-fdf43aa6f949-ceph\") pod \"cinder-backup-0\" (UID: \"caa349ef-0de7-46c4-a7cd-fdf43aa6f949\") " pod="openstack/cinder-backup-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.874507 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7d74t\" (UniqueName: \"kubernetes.io/projected/caa349ef-0de7-46c4-a7cd-fdf43aa6f949-kube-api-access-7d74t\") pod \"cinder-backup-0\" (UID: \"caa349ef-0de7-46c4-a7cd-fdf43aa6f949\") " pod="openstack/cinder-backup-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.939970 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:42 crc kubenswrapper[4799]: I1129 05:28:42.941744 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.056362 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-create-r7r7q"] Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.058251 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-r7r7q" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.083635 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-r7r7q"] Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.111536 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-566ff59745-jp249"] Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.113459 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-566ff59745-jp249" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.117273 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.120078 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.126471 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-pg86b" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.126616 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.173310 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vk8wt\" (UniqueName: \"kubernetes.io/projected/9e368006-30f3-4355-86ca-7ac3ec9cbbe7-kube-api-access-vk8wt\") pod \"manila-db-create-r7r7q\" (UID: \"9e368006-30f3-4355-86ca-7ac3ec9cbbe7\") " pod="openstack/manila-db-create-r7r7q" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.173704 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9e368006-30f3-4355-86ca-7ac3ec9cbbe7-operator-scripts\") pod \"manila-db-create-r7r7q\" (UID: \"9e368006-30f3-4355-86ca-7ac3ec9cbbe7\") " pod="openstack/manila-db-create-r7r7q" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.187736 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-566ff59745-jp249"] Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.204522 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-787d-account-create-update-9qmsv"] Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.206704 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-787d-account-create-update-9qmsv" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.224564 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-db-secret" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.247551 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-787d-account-create-update-9qmsv"] Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.282247 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ebcbc04a-230e-46b1-be0d-af9ece769369-logs\") pod \"horizon-566ff59745-jp249\" (UID: \"ebcbc04a-230e-46b1-be0d-af9ece769369\") " pod="openstack/horizon-566ff59745-jp249" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.282318 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ebcbc04a-230e-46b1-be0d-af9ece769369-horizon-secret-key\") pod \"horizon-566ff59745-jp249\" (UID: \"ebcbc04a-230e-46b1-be0d-af9ece769369\") " pod="openstack/horizon-566ff59745-jp249" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.282348 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ebcbc04a-230e-46b1-be0d-af9ece769369-scripts\") pod \"horizon-566ff59745-jp249\" (UID: \"ebcbc04a-230e-46b1-be0d-af9ece769369\") " pod="openstack/horizon-566ff59745-jp249" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.282389 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ls4wl\" (UniqueName: \"kubernetes.io/projected/ebcbc04a-230e-46b1-be0d-af9ece769369-kube-api-access-ls4wl\") pod \"horizon-566ff59745-jp249\" (UID: \"ebcbc04a-230e-46b1-be0d-af9ece769369\") " pod="openstack/horizon-566ff59745-jp249" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.282512 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vk8wt\" (UniqueName: \"kubernetes.io/projected/9e368006-30f3-4355-86ca-7ac3ec9cbbe7-kube-api-access-vk8wt\") pod \"manila-db-create-r7r7q\" (UID: \"9e368006-30f3-4355-86ca-7ac3ec9cbbe7\") " pod="openstack/manila-db-create-r7r7q" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.282536 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9e368006-30f3-4355-86ca-7ac3ec9cbbe7-operator-scripts\") pod \"manila-db-create-r7r7q\" (UID: \"9e368006-30f3-4355-86ca-7ac3ec9cbbe7\") " pod="openstack/manila-db-create-r7r7q" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.282587 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ebcbc04a-230e-46b1-be0d-af9ece769369-config-data\") pod \"horizon-566ff59745-jp249\" (UID: \"ebcbc04a-230e-46b1-be0d-af9ece769369\") " pod="openstack/horizon-566ff59745-jp249" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.284266 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9e368006-30f3-4355-86ca-7ac3ec9cbbe7-operator-scripts\") pod \"manila-db-create-r7r7q\" (UID: \"9e368006-30f3-4355-86ca-7ac3ec9cbbe7\") " pod="openstack/manila-db-create-r7r7q" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.318485 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vk8wt\" (UniqueName: \"kubernetes.io/projected/9e368006-30f3-4355-86ca-7ac3ec9cbbe7-kube-api-access-vk8wt\") pod \"manila-db-create-r7r7q\" (UID: \"9e368006-30f3-4355-86ca-7ac3ec9cbbe7\") " pod="openstack/manila-db-create-r7r7q" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.339828 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.341836 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.350250 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.350270 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.350679 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.350732 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-fhqdz" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.361695 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-598444b44c-v5znr"] Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.364166 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-598444b44c-v5znr" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.385899 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ebcbc04a-230e-46b1-be0d-af9ece769369-config-data\") pod \"horizon-566ff59745-jp249\" (UID: \"ebcbc04a-230e-46b1-be0d-af9ece769369\") " pod="openstack/horizon-566ff59745-jp249" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.385989 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ebcbc04a-230e-46b1-be0d-af9ece769369-logs\") pod \"horizon-566ff59745-jp249\" (UID: \"ebcbc04a-230e-46b1-be0d-af9ece769369\") " pod="openstack/horizon-566ff59745-jp249" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.386028 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ebcbc04a-230e-46b1-be0d-af9ece769369-horizon-secret-key\") pod \"horizon-566ff59745-jp249\" (UID: \"ebcbc04a-230e-46b1-be0d-af9ece769369\") " pod="openstack/horizon-566ff59745-jp249" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.386047 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ebcbc04a-230e-46b1-be0d-af9ece769369-scripts\") pod \"horizon-566ff59745-jp249\" (UID: \"ebcbc04a-230e-46b1-be0d-af9ece769369\") " pod="openstack/horizon-566ff59745-jp249" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.386079 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ls4wl\" (UniqueName: \"kubernetes.io/projected/ebcbc04a-230e-46b1-be0d-af9ece769369-kube-api-access-ls4wl\") pod \"horizon-566ff59745-jp249\" (UID: \"ebcbc04a-230e-46b1-be0d-af9ece769369\") " pod="openstack/horizon-566ff59745-jp249" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.386121 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88d65\" (UniqueName: \"kubernetes.io/projected/81148106-ce19-4be8-8afd-0db4518977b0-kube-api-access-88d65\") pod \"manila-787d-account-create-update-9qmsv\" (UID: \"81148106-ce19-4be8-8afd-0db4518977b0\") " pod="openstack/manila-787d-account-create-update-9qmsv" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.386163 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81148106-ce19-4be8-8afd-0db4518977b0-operator-scripts\") pod \"manila-787d-account-create-update-9qmsv\" (UID: \"81148106-ce19-4be8-8afd-0db4518977b0\") " pod="openstack/manila-787d-account-create-update-9qmsv" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.387019 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ebcbc04a-230e-46b1-be0d-af9ece769369-scripts\") pod \"horizon-566ff59745-jp249\" (UID: \"ebcbc04a-230e-46b1-be0d-af9ece769369\") " pod="openstack/horizon-566ff59745-jp249" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.387137 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ebcbc04a-230e-46b1-be0d-af9ece769369-logs\") pod \"horizon-566ff59745-jp249\" (UID: \"ebcbc04a-230e-46b1-be0d-af9ece769369\") " pod="openstack/horizon-566ff59745-jp249" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.388107 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-r7r7q" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.391514 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.391883 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ebcbc04a-230e-46b1-be0d-af9ece769369-horizon-secret-key\") pod \"horizon-566ff59745-jp249\" (UID: \"ebcbc04a-230e-46b1-be0d-af9ece769369\") " pod="openstack/horizon-566ff59745-jp249" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.394155 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ebcbc04a-230e-46b1-be0d-af9ece769369-config-data\") pod \"horizon-566ff59745-jp249\" (UID: \"ebcbc04a-230e-46b1-be0d-af9ece769369\") " pod="openstack/horizon-566ff59745-jp249" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.414413 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ls4wl\" (UniqueName: \"kubernetes.io/projected/ebcbc04a-230e-46b1-be0d-af9ece769369-kube-api-access-ls4wl\") pod \"horizon-566ff59745-jp249\" (UID: \"ebcbc04a-230e-46b1-be0d-af9ece769369\") " pod="openstack/horizon-566ff59745-jp249" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.419752 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-598444b44c-v5znr"] Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.437337 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-566ff59745-jp249" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.448563 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.452761 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.455704 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.456174 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.462058 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.491008 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67c0f919-41ec-4c54-a8d4-fea37e8c3efb-config-data\") pod \"glance-default-external-api-0\" (UID: \"67c0f919-41ec-4c54-a8d4-fea37e8c3efb\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.491056 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/67c0f919-41ec-4c54-a8d4-fea37e8c3efb-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"67c0f919-41ec-4c54-a8d4-fea37e8c3efb\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.491081 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfqfv\" (UniqueName: \"kubernetes.io/projected/67c0f919-41ec-4c54-a8d4-fea37e8c3efb-kube-api-access-dfqfv\") pod \"glance-default-external-api-0\" (UID: \"67c0f919-41ec-4c54-a8d4-fea37e8c3efb\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.491114 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81148106-ce19-4be8-8afd-0db4518977b0-operator-scripts\") pod \"manila-787d-account-create-update-9qmsv\" (UID: \"81148106-ce19-4be8-8afd-0db4518977b0\") " pod="openstack/manila-787d-account-create-update-9qmsv" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.491287 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67c0f919-41ec-4c54-a8d4-fea37e8c3efb-logs\") pod \"glance-default-external-api-0\" (UID: \"67c0f919-41ec-4c54-a8d4-fea37e8c3efb\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.491332 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/67c0f919-41ec-4c54-a8d4-fea37e8c3efb-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"67c0f919-41ec-4c54-a8d4-fea37e8c3efb\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.491502 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rzbx\" (UniqueName: \"kubernetes.io/projected/04de4a38-5d09-4c68-b7a2-be6482748e50-kube-api-access-8rzbx\") pod \"horizon-598444b44c-v5znr\" (UID: \"04de4a38-5d09-4c68-b7a2-be6482748e50\") " pod="openstack/horizon-598444b44c-v5znr" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.491612 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67c0f919-41ec-4c54-a8d4-fea37e8c3efb-scripts\") pod \"glance-default-external-api-0\" (UID: \"67c0f919-41ec-4c54-a8d4-fea37e8c3efb\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.491647 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67c0f919-41ec-4c54-a8d4-fea37e8c3efb-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"67c0f919-41ec-4c54-a8d4-fea37e8c3efb\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.491698 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/04de4a38-5d09-4c68-b7a2-be6482748e50-horizon-secret-key\") pod \"horizon-598444b44c-v5znr\" (UID: \"04de4a38-5d09-4c68-b7a2-be6482748e50\") " pod="openstack/horizon-598444b44c-v5znr" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.491736 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04de4a38-5d09-4c68-b7a2-be6482748e50-logs\") pod \"horizon-598444b44c-v5znr\" (UID: \"04de4a38-5d09-4c68-b7a2-be6482748e50\") " pod="openstack/horizon-598444b44c-v5znr" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.491801 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/04de4a38-5d09-4c68-b7a2-be6482748e50-config-data\") pod \"horizon-598444b44c-v5znr\" (UID: \"04de4a38-5d09-4c68-b7a2-be6482748e50\") " pod="openstack/horizon-598444b44c-v5znr" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.491826 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81148106-ce19-4be8-8afd-0db4518977b0-operator-scripts\") pod \"manila-787d-account-create-update-9qmsv\" (UID: \"81148106-ce19-4be8-8afd-0db4518977b0\") " pod="openstack/manila-787d-account-create-update-9qmsv" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.491916 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/67c0f919-41ec-4c54-a8d4-fea37e8c3efb-ceph\") pod \"glance-default-external-api-0\" (UID: \"67c0f919-41ec-4c54-a8d4-fea37e8c3efb\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.492005 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/04de4a38-5d09-4c68-b7a2-be6482748e50-scripts\") pod \"horizon-598444b44c-v5znr\" (UID: \"04de4a38-5d09-4c68-b7a2-be6482748e50\") " pod="openstack/horizon-598444b44c-v5znr" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.492300 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"67c0f919-41ec-4c54-a8d4-fea37e8c3efb\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.492362 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88d65\" (UniqueName: \"kubernetes.io/projected/81148106-ce19-4be8-8afd-0db4518977b0-kube-api-access-88d65\") pod \"manila-787d-account-create-update-9qmsv\" (UID: \"81148106-ce19-4be8-8afd-0db4518977b0\") " pod="openstack/manila-787d-account-create-update-9qmsv" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.517638 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88d65\" (UniqueName: \"kubernetes.io/projected/81148106-ce19-4be8-8afd-0db4518977b0-kube-api-access-88d65\") pod \"manila-787d-account-create-update-9qmsv\" (UID: \"81148106-ce19-4be8-8afd-0db4518977b0\") " pod="openstack/manila-787d-account-create-update-9qmsv" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.547486 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-787d-account-create-update-9qmsv" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.594284 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04de4a38-5d09-4c68-b7a2-be6482748e50-logs\") pod \"horizon-598444b44c-v5znr\" (UID: \"04de4a38-5d09-4c68-b7a2-be6482748e50\") " pod="openstack/horizon-598444b44c-v5znr" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.594338 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/04de4a38-5d09-4c68-b7a2-be6482748e50-config-data\") pod \"horizon-598444b44c-v5znr\" (UID: \"04de4a38-5d09-4c68-b7a2-be6482748e50\") " pod="openstack/horizon-598444b44c-v5znr" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.594380 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/304c844f-62a9-4253-b68b-a5003341cea1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"304c844f-62a9-4253-b68b-a5003341cea1\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.594405 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/304c844f-62a9-4253-b68b-a5003341cea1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"304c844f-62a9-4253-b68b-a5003341cea1\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.594425 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/304c844f-62a9-4253-b68b-a5003341cea1-logs\") pod \"glance-default-internal-api-0\" (UID: \"304c844f-62a9-4253-b68b-a5003341cea1\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.594449 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/67c0f919-41ec-4c54-a8d4-fea37e8c3efb-ceph\") pod \"glance-default-external-api-0\" (UID: \"67c0f919-41ec-4c54-a8d4-fea37e8c3efb\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.594470 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jdss\" (UniqueName: \"kubernetes.io/projected/304c844f-62a9-4253-b68b-a5003341cea1-kube-api-access-8jdss\") pod \"glance-default-internal-api-0\" (UID: \"304c844f-62a9-4253-b68b-a5003341cea1\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.594507 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/04de4a38-5d09-4c68-b7a2-be6482748e50-scripts\") pod \"horizon-598444b44c-v5znr\" (UID: \"04de4a38-5d09-4c68-b7a2-be6482748e50\") " pod="openstack/horizon-598444b44c-v5znr" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.594560 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"67c0f919-41ec-4c54-a8d4-fea37e8c3efb\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.594593 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67c0f919-41ec-4c54-a8d4-fea37e8c3efb-config-data\") pod \"glance-default-external-api-0\" (UID: \"67c0f919-41ec-4c54-a8d4-fea37e8c3efb\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.594613 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/67c0f919-41ec-4c54-a8d4-fea37e8c3efb-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"67c0f919-41ec-4c54-a8d4-fea37e8c3efb\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.594644 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfqfv\" (UniqueName: \"kubernetes.io/projected/67c0f919-41ec-4c54-a8d4-fea37e8c3efb-kube-api-access-dfqfv\") pod \"glance-default-external-api-0\" (UID: \"67c0f919-41ec-4c54-a8d4-fea37e8c3efb\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.594673 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/304c844f-62a9-4253-b68b-a5003341cea1-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"304c844f-62a9-4253-b68b-a5003341cea1\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.594697 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67c0f919-41ec-4c54-a8d4-fea37e8c3efb-logs\") pod \"glance-default-external-api-0\" (UID: \"67c0f919-41ec-4c54-a8d4-fea37e8c3efb\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.594711 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/67c0f919-41ec-4c54-a8d4-fea37e8c3efb-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"67c0f919-41ec-4c54-a8d4-fea37e8c3efb\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.594748 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/304c844f-62a9-4253-b68b-a5003341cea1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"304c844f-62a9-4253-b68b-a5003341cea1\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.594766 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rzbx\" (UniqueName: \"kubernetes.io/projected/04de4a38-5d09-4c68-b7a2-be6482748e50-kube-api-access-8rzbx\") pod \"horizon-598444b44c-v5znr\" (UID: \"04de4a38-5d09-4c68-b7a2-be6482748e50\") " pod="openstack/horizon-598444b44c-v5znr" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.594817 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"304c844f-62a9-4253-b68b-a5003341cea1\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.594839 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/304c844f-62a9-4253-b68b-a5003341cea1-ceph\") pod \"glance-default-internal-api-0\" (UID: \"304c844f-62a9-4253-b68b-a5003341cea1\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.594866 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67c0f919-41ec-4c54-a8d4-fea37e8c3efb-scripts\") pod \"glance-default-external-api-0\" (UID: \"67c0f919-41ec-4c54-a8d4-fea37e8c3efb\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.594885 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67c0f919-41ec-4c54-a8d4-fea37e8c3efb-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"67c0f919-41ec-4c54-a8d4-fea37e8c3efb\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.594909 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/304c844f-62a9-4253-b68b-a5003341cea1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"304c844f-62a9-4253-b68b-a5003341cea1\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.594928 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/04de4a38-5d09-4c68-b7a2-be6482748e50-horizon-secret-key\") pod \"horizon-598444b44c-v5znr\" (UID: \"04de4a38-5d09-4c68-b7a2-be6482748e50\") " pod="openstack/horizon-598444b44c-v5znr" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.596727 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04de4a38-5d09-4c68-b7a2-be6482748e50-logs\") pod \"horizon-598444b44c-v5znr\" (UID: \"04de4a38-5d09-4c68-b7a2-be6482748e50\") " pod="openstack/horizon-598444b44c-v5znr" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.599471 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/04de4a38-5d09-4c68-b7a2-be6482748e50-scripts\") pod \"horizon-598444b44c-v5znr\" (UID: \"04de4a38-5d09-4c68-b7a2-be6482748e50\") " pod="openstack/horizon-598444b44c-v5znr" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.600272 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"67c0f919-41ec-4c54-a8d4-fea37e8c3efb\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.602265 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/67c0f919-41ec-4c54-a8d4-fea37e8c3efb-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"67c0f919-41ec-4c54-a8d4-fea37e8c3efb\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.603474 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67c0f919-41ec-4c54-a8d4-fea37e8c3efb-logs\") pod \"glance-default-external-api-0\" (UID: \"67c0f919-41ec-4c54-a8d4-fea37e8c3efb\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.603738 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/04de4a38-5d09-4c68-b7a2-be6482748e50-horizon-secret-key\") pod \"horizon-598444b44c-v5znr\" (UID: \"04de4a38-5d09-4c68-b7a2-be6482748e50\") " pod="openstack/horizon-598444b44c-v5znr" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.606941 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/67c0f919-41ec-4c54-a8d4-fea37e8c3efb-ceph\") pod \"glance-default-external-api-0\" (UID: \"67c0f919-41ec-4c54-a8d4-fea37e8c3efb\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.610241 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/67c0f919-41ec-4c54-a8d4-fea37e8c3efb-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"67c0f919-41ec-4c54-a8d4-fea37e8c3efb\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.613716 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67c0f919-41ec-4c54-a8d4-fea37e8c3efb-scripts\") pod \"glance-default-external-api-0\" (UID: \"67c0f919-41ec-4c54-a8d4-fea37e8c3efb\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.614137 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67c0f919-41ec-4c54-a8d4-fea37e8c3efb-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"67c0f919-41ec-4c54-a8d4-fea37e8c3efb\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.620452 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rzbx\" (UniqueName: \"kubernetes.io/projected/04de4a38-5d09-4c68-b7a2-be6482748e50-kube-api-access-8rzbx\") pod \"horizon-598444b44c-v5znr\" (UID: \"04de4a38-5d09-4c68-b7a2-be6482748e50\") " pod="openstack/horizon-598444b44c-v5znr" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.626528 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfqfv\" (UniqueName: \"kubernetes.io/projected/67c0f919-41ec-4c54-a8d4-fea37e8c3efb-kube-api-access-dfqfv\") pod \"glance-default-external-api-0\" (UID: \"67c0f919-41ec-4c54-a8d4-fea37e8c3efb\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.627934 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67c0f919-41ec-4c54-a8d4-fea37e8c3efb-config-data\") pod \"glance-default-external-api-0\" (UID: \"67c0f919-41ec-4c54-a8d4-fea37e8c3efb\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.629404 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/04de4a38-5d09-4c68-b7a2-be6482748e50-config-data\") pod \"horizon-598444b44c-v5znr\" (UID: \"04de4a38-5d09-4c68-b7a2-be6482748e50\") " pod="openstack/horizon-598444b44c-v5znr" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.660970 4799 scope.go:117] "RemoveContainer" containerID="e6d4b3d1d5b5e1eeb13508d67de3d3c3aac52e76337b55e53793185c0d001d82" Nov 29 05:28:43 crc kubenswrapper[4799]: E1129 05:28:43.661182 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.676519 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"67c0f919-41ec-4c54-a8d4-fea37e8c3efb\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.682130 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.691707 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-598444b44c-v5znr" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.696887 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/304c844f-62a9-4253-b68b-a5003341cea1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"304c844f-62a9-4253-b68b-a5003341cea1\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.696937 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/304c844f-62a9-4253-b68b-a5003341cea1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"304c844f-62a9-4253-b68b-a5003341cea1\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.696959 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/304c844f-62a9-4253-b68b-a5003341cea1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"304c844f-62a9-4253-b68b-a5003341cea1\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.696977 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/304c844f-62a9-4253-b68b-a5003341cea1-logs\") pod \"glance-default-internal-api-0\" (UID: \"304c844f-62a9-4253-b68b-a5003341cea1\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.696997 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jdss\" (UniqueName: \"kubernetes.io/projected/304c844f-62a9-4253-b68b-a5003341cea1-kube-api-access-8jdss\") pod \"glance-default-internal-api-0\" (UID: \"304c844f-62a9-4253-b68b-a5003341cea1\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.697077 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/304c844f-62a9-4253-b68b-a5003341cea1-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"304c844f-62a9-4253-b68b-a5003341cea1\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.697119 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/304c844f-62a9-4253-b68b-a5003341cea1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"304c844f-62a9-4253-b68b-a5003341cea1\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.697142 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"304c844f-62a9-4253-b68b-a5003341cea1\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.697159 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/304c844f-62a9-4253-b68b-a5003341cea1-ceph\") pod \"glance-default-internal-api-0\" (UID: \"304c844f-62a9-4253-b68b-a5003341cea1\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.698862 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/304c844f-62a9-4253-b68b-a5003341cea1-logs\") pod \"glance-default-internal-api-0\" (UID: \"304c844f-62a9-4253-b68b-a5003341cea1\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.699448 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/304c844f-62a9-4253-b68b-a5003341cea1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"304c844f-62a9-4253-b68b-a5003341cea1\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.699632 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"304c844f-62a9-4253-b68b-a5003341cea1\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.702402 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/304c844f-62a9-4253-b68b-a5003341cea1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"304c844f-62a9-4253-b68b-a5003341cea1\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.704428 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/304c844f-62a9-4253-b68b-a5003341cea1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"304c844f-62a9-4253-b68b-a5003341cea1\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.713128 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/304c844f-62a9-4253-b68b-a5003341cea1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"304c844f-62a9-4253-b68b-a5003341cea1\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.718767 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/304c844f-62a9-4253-b68b-a5003341cea1-ceph\") pod \"glance-default-internal-api-0\" (UID: \"304c844f-62a9-4253-b68b-a5003341cea1\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.721534 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/304c844f-62a9-4253-b68b-a5003341cea1-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"304c844f-62a9-4253-b68b-a5003341cea1\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.746644 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jdss\" (UniqueName: \"kubernetes.io/projected/304c844f-62a9-4253-b68b-a5003341cea1-kube-api-access-8jdss\") pod \"glance-default-internal-api-0\" (UID: \"304c844f-62a9-4253-b68b-a5003341cea1\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.783761 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.809275 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"304c844f-62a9-4253-b68b-a5003341cea1\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.921510 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Nov 29 05:28:43 crc kubenswrapper[4799]: W1129 05:28:43.938656 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0907967b_0bf5_4fe0_9862_ffbd54efde7f.slice/crio-c65ccf751b58971d2441c44939b1044013a3a0047d2813da38d4f5eda5cd8e38 WatchSource:0}: Error finding container c65ccf751b58971d2441c44939b1044013a3a0047d2813da38d4f5eda5cd8e38: Status 404 returned error can't find the container with id c65ccf751b58971d2441c44939b1044013a3a0047d2813da38d4f5eda5cd8e38 Nov 29 05:28:43 crc kubenswrapper[4799]: I1129 05:28:43.959959 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"caa349ef-0de7-46c4-a7cd-fdf43aa6f949","Type":"ContainerStarted","Data":"bfcf54c97c9ee8af778e24428fc7c98395677540bbcc5d7c43f2e2cf5215bd28"} Nov 29 05:28:44 crc kubenswrapper[4799]: I1129 05:28:44.000677 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-r7r7q"] Nov 29 05:28:44 crc kubenswrapper[4799]: W1129 05:28:44.001656 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9e368006_30f3_4355_86ca_7ac3ec9cbbe7.slice/crio-444d4532ff7e7dd16652742240dec3a3f6067821fdcc9eefb3b7fbd1f78408ca WatchSource:0}: Error finding container 444d4532ff7e7dd16652742240dec3a3f6067821fdcc9eefb3b7fbd1f78408ca: Status 404 returned error can't find the container with id 444d4532ff7e7dd16652742240dec3a3f6067821fdcc9eefb3b7fbd1f78408ca Nov 29 05:28:44 crc kubenswrapper[4799]: I1129 05:28:44.096722 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 29 05:28:44 crc kubenswrapper[4799]: I1129 05:28:44.101617 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-566ff59745-jp249"] Nov 29 05:28:44 crc kubenswrapper[4799]: I1129 05:28:44.202372 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-787d-account-create-update-9qmsv"] Nov 29 05:28:44 crc kubenswrapper[4799]: W1129 05:28:44.215629 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod81148106_ce19_4be8_8afd_0db4518977b0.slice/crio-a084ef8b4e965a0331d8e283f5c16b8744629a6fb2ea9c628a186b6e2669d147 WatchSource:0}: Error finding container a084ef8b4e965a0331d8e283f5c16b8744629a6fb2ea9c628a186b6e2669d147: Status 404 returned error can't find the container with id a084ef8b4e965a0331d8e283f5c16b8744629a6fb2ea9c628a186b6e2669d147 Nov 29 05:28:44 crc kubenswrapper[4799]: I1129 05:28:44.395413 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-598444b44c-v5znr"] Nov 29 05:28:44 crc kubenswrapper[4799]: I1129 05:28:44.468107 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 29 05:28:44 crc kubenswrapper[4799]: I1129 05:28:44.687437 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 29 05:28:44 crc kubenswrapper[4799]: W1129 05:28:44.697031 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod304c844f_62a9_4253_b68b_a5003341cea1.slice/crio-96247213ded49154e66366af087376955e1b6cae935b311ca1bd731d50477a4b WatchSource:0}: Error finding container 96247213ded49154e66366af087376955e1b6cae935b311ca1bd731d50477a4b: Status 404 returned error can't find the container with id 96247213ded49154e66366af087376955e1b6cae935b311ca1bd731d50477a4b Nov 29 05:28:44 crc kubenswrapper[4799]: I1129 05:28:44.990657 4799 generic.go:334] "Generic (PLEG): container finished" podID="81148106-ce19-4be8-8afd-0db4518977b0" containerID="a1d3cd6cc4df20687edb72bd63e3aaa1e58621a04c9675d0f66617bb551cdcb4" exitCode=0 Nov 29 05:28:44 crc kubenswrapper[4799]: I1129 05:28:44.991540 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-787d-account-create-update-9qmsv" event={"ID":"81148106-ce19-4be8-8afd-0db4518977b0","Type":"ContainerDied","Data":"a1d3cd6cc4df20687edb72bd63e3aaa1e58621a04c9675d0f66617bb551cdcb4"} Nov 29 05:28:44 crc kubenswrapper[4799]: I1129 05:28:44.991575 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-787d-account-create-update-9qmsv" event={"ID":"81148106-ce19-4be8-8afd-0db4518977b0","Type":"ContainerStarted","Data":"a084ef8b4e965a0331d8e283f5c16b8744629a6fb2ea9c628a186b6e2669d147"} Nov 29 05:28:44 crc kubenswrapper[4799]: I1129 05:28:44.994999 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"304c844f-62a9-4253-b68b-a5003341cea1","Type":"ContainerStarted","Data":"96247213ded49154e66366af087376955e1b6cae935b311ca1bd731d50477a4b"} Nov 29 05:28:45 crc kubenswrapper[4799]: I1129 05:28:45.001234 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-566ff59745-jp249" event={"ID":"ebcbc04a-230e-46b1-be0d-af9ece769369","Type":"ContainerStarted","Data":"b7b41b27a0067adde7d22f2cc2d1916b9a2cfbbfcc3eeb4732115a1442f63e6a"} Nov 29 05:28:45 crc kubenswrapper[4799]: I1129 05:28:45.007282 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"67c0f919-41ec-4c54-a8d4-fea37e8c3efb","Type":"ContainerStarted","Data":"b90cad992bc106407343de0b2c52191e30df994ae3c54a07fed012509b688d95"} Nov 29 05:28:45 crc kubenswrapper[4799]: I1129 05:28:45.015013 4799 generic.go:334] "Generic (PLEG): container finished" podID="9e368006-30f3-4355-86ca-7ac3ec9cbbe7" containerID="bea20aa939d744e48585ce7b6a148d714c9d1ff15e8300f3fc8c0120a42fbc9c" exitCode=0 Nov 29 05:28:45 crc kubenswrapper[4799]: I1129 05:28:45.015580 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-r7r7q" event={"ID":"9e368006-30f3-4355-86ca-7ac3ec9cbbe7","Type":"ContainerDied","Data":"bea20aa939d744e48585ce7b6a148d714c9d1ff15e8300f3fc8c0120a42fbc9c"} Nov 29 05:28:45 crc kubenswrapper[4799]: I1129 05:28:45.015611 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-r7r7q" event={"ID":"9e368006-30f3-4355-86ca-7ac3ec9cbbe7","Type":"ContainerStarted","Data":"444d4532ff7e7dd16652742240dec3a3f6067821fdcc9eefb3b7fbd1f78408ca"} Nov 29 05:28:45 crc kubenswrapper[4799]: I1129 05:28:45.017271 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"0907967b-0bf5-4fe0-9862-ffbd54efde7f","Type":"ContainerStarted","Data":"c65ccf751b58971d2441c44939b1044013a3a0047d2813da38d4f5eda5cd8e38"} Nov 29 05:28:45 crc kubenswrapper[4799]: I1129 05:28:45.029556 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-598444b44c-v5znr" event={"ID":"04de4a38-5d09-4c68-b7a2-be6482748e50","Type":"ContainerStarted","Data":"af552d396b39049c076f4bbbd89611106fca62ea40b2ec616195df247cba534e"} Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.060185 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"304c844f-62a9-4253-b68b-a5003341cea1","Type":"ContainerStarted","Data":"aaf55259341ee3c336f469d5eae5425cff1a95608ba68c660aa005ba5ed84d2d"} Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.065076 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"67c0f919-41ec-4c54-a8d4-fea37e8c3efb","Type":"ContainerStarted","Data":"d20988a5ceb601b7bf3e1991eca4dad35d9042054d41d5047aabed0d25b30acc"} Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.069180 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"0907967b-0bf5-4fe0-9862-ffbd54efde7f","Type":"ContainerStarted","Data":"9d7a26425879e29cfc7dbd73cdf52f5c71f55dabe72d42505030bc62dda93f51"} Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.303753 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-566ff59745-jp249"] Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.368850 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.391371 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-86d9b546cb-5xxhb"] Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.393453 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-86d9b546cb-5xxhb" Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.396943 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.412164 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-86d9b546cb-5xxhb"] Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.490824 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.523328 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f09087c5-87d5-4cd1-9a91-cbe48713d9e9-combined-ca-bundle\") pod \"horizon-86d9b546cb-5xxhb\" (UID: \"f09087c5-87d5-4cd1-9a91-cbe48713d9e9\") " pod="openstack/horizon-86d9b546cb-5xxhb" Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.523393 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f09087c5-87d5-4cd1-9a91-cbe48713d9e9-horizon-secret-key\") pod \"horizon-86d9b546cb-5xxhb\" (UID: \"f09087c5-87d5-4cd1-9a91-cbe48713d9e9\") " pod="openstack/horizon-86d9b546cb-5xxhb" Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.523416 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f09087c5-87d5-4cd1-9a91-cbe48713d9e9-logs\") pod \"horizon-86d9b546cb-5xxhb\" (UID: \"f09087c5-87d5-4cd1-9a91-cbe48713d9e9\") " pod="openstack/horizon-86d9b546cb-5xxhb" Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.523527 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f09087c5-87d5-4cd1-9a91-cbe48713d9e9-config-data\") pod \"horizon-86d9b546cb-5xxhb\" (UID: \"f09087c5-87d5-4cd1-9a91-cbe48713d9e9\") " pod="openstack/horizon-86d9b546cb-5xxhb" Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.523565 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/f09087c5-87d5-4cd1-9a91-cbe48713d9e9-horizon-tls-certs\") pod \"horizon-86d9b546cb-5xxhb\" (UID: \"f09087c5-87d5-4cd1-9a91-cbe48713d9e9\") " pod="openstack/horizon-86d9b546cb-5xxhb" Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.523626 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4b5g\" (UniqueName: \"kubernetes.io/projected/f09087c5-87d5-4cd1-9a91-cbe48713d9e9-kube-api-access-p4b5g\") pod \"horizon-86d9b546cb-5xxhb\" (UID: \"f09087c5-87d5-4cd1-9a91-cbe48713d9e9\") " pod="openstack/horizon-86d9b546cb-5xxhb" Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.523653 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f09087c5-87d5-4cd1-9a91-cbe48713d9e9-scripts\") pod \"horizon-86d9b546cb-5xxhb\" (UID: \"f09087c5-87d5-4cd1-9a91-cbe48713d9e9\") " pod="openstack/horizon-86d9b546cb-5xxhb" Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.572858 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-598444b44c-v5znr"] Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.590807 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7c5488f496-jvw7p"] Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.592739 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c5488f496-jvw7p" Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.598850 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7c5488f496-jvw7p"] Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.644334 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/93fb8bea-4246-40da-b480-c980b628e2b9-config-data\") pod \"horizon-7c5488f496-jvw7p\" (UID: \"93fb8bea-4246-40da-b480-c980b628e2b9\") " pod="openstack/horizon-7c5488f496-jvw7p" Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.644758 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f09087c5-87d5-4cd1-9a91-cbe48713d9e9-combined-ca-bundle\") pod \"horizon-86d9b546cb-5xxhb\" (UID: \"f09087c5-87d5-4cd1-9a91-cbe48713d9e9\") " pod="openstack/horizon-86d9b546cb-5xxhb" Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.645157 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f09087c5-87d5-4cd1-9a91-cbe48713d9e9-horizon-secret-key\") pod \"horizon-86d9b546cb-5xxhb\" (UID: \"f09087c5-87d5-4cd1-9a91-cbe48713d9e9\") " pod="openstack/horizon-86d9b546cb-5xxhb" Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.645256 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f09087c5-87d5-4cd1-9a91-cbe48713d9e9-logs\") pod \"horizon-86d9b546cb-5xxhb\" (UID: \"f09087c5-87d5-4cd1-9a91-cbe48713d9e9\") " pod="openstack/horizon-86d9b546cb-5xxhb" Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.645421 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93fb8bea-4246-40da-b480-c980b628e2b9-combined-ca-bundle\") pod \"horizon-7c5488f496-jvw7p\" (UID: \"93fb8bea-4246-40da-b480-c980b628e2b9\") " pod="openstack/horizon-7c5488f496-jvw7p" Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.645601 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/93fb8bea-4246-40da-b480-c980b628e2b9-scripts\") pod \"horizon-7c5488f496-jvw7p\" (UID: \"93fb8bea-4246-40da-b480-c980b628e2b9\") " pod="openstack/horizon-7c5488f496-jvw7p" Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.645694 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f09087c5-87d5-4cd1-9a91-cbe48713d9e9-config-data\") pod \"horizon-86d9b546cb-5xxhb\" (UID: \"f09087c5-87d5-4cd1-9a91-cbe48713d9e9\") " pod="openstack/horizon-86d9b546cb-5xxhb" Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.645782 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/f09087c5-87d5-4cd1-9a91-cbe48713d9e9-horizon-tls-certs\") pod \"horizon-86d9b546cb-5xxhb\" (UID: \"f09087c5-87d5-4cd1-9a91-cbe48713d9e9\") " pod="openstack/horizon-86d9b546cb-5xxhb" Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.645925 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4b5g\" (UniqueName: \"kubernetes.io/projected/f09087c5-87d5-4cd1-9a91-cbe48713d9e9-kube-api-access-p4b5g\") pod \"horizon-86d9b546cb-5xxhb\" (UID: \"f09087c5-87d5-4cd1-9a91-cbe48713d9e9\") " pod="openstack/horizon-86d9b546cb-5xxhb" Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.646003 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f09087c5-87d5-4cd1-9a91-cbe48713d9e9-scripts\") pod \"horizon-86d9b546cb-5xxhb\" (UID: \"f09087c5-87d5-4cd1-9a91-cbe48713d9e9\") " pod="openstack/horizon-86d9b546cb-5xxhb" Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.649939 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/93fb8bea-4246-40da-b480-c980b628e2b9-horizon-tls-certs\") pod \"horizon-7c5488f496-jvw7p\" (UID: \"93fb8bea-4246-40da-b480-c980b628e2b9\") " pod="openstack/horizon-7c5488f496-jvw7p" Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.650068 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/93fb8bea-4246-40da-b480-c980b628e2b9-horizon-secret-key\") pod \"horizon-7c5488f496-jvw7p\" (UID: \"93fb8bea-4246-40da-b480-c980b628e2b9\") " pod="openstack/horizon-7c5488f496-jvw7p" Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.650158 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/93fb8bea-4246-40da-b480-c980b628e2b9-logs\") pod \"horizon-7c5488f496-jvw7p\" (UID: \"93fb8bea-4246-40da-b480-c980b628e2b9\") " pod="openstack/horizon-7c5488f496-jvw7p" Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.650284 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5rmc\" (UniqueName: \"kubernetes.io/projected/93fb8bea-4246-40da-b480-c980b628e2b9-kube-api-access-p5rmc\") pod \"horizon-7c5488f496-jvw7p\" (UID: \"93fb8bea-4246-40da-b480-c980b628e2b9\") " pod="openstack/horizon-7c5488f496-jvw7p" Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.655107 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f09087c5-87d5-4cd1-9a91-cbe48713d9e9-scripts\") pod \"horizon-86d9b546cb-5xxhb\" (UID: \"f09087c5-87d5-4cd1-9a91-cbe48713d9e9\") " pod="openstack/horizon-86d9b546cb-5xxhb" Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.655361 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f09087c5-87d5-4cd1-9a91-cbe48713d9e9-config-data\") pod \"horizon-86d9b546cb-5xxhb\" (UID: \"f09087c5-87d5-4cd1-9a91-cbe48713d9e9\") " pod="openstack/horizon-86d9b546cb-5xxhb" Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.656411 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f09087c5-87d5-4cd1-9a91-cbe48713d9e9-logs\") pod \"horizon-86d9b546cb-5xxhb\" (UID: \"f09087c5-87d5-4cd1-9a91-cbe48713d9e9\") " pod="openstack/horizon-86d9b546cb-5xxhb" Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.680481 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f09087c5-87d5-4cd1-9a91-cbe48713d9e9-horizon-secret-key\") pod \"horizon-86d9b546cb-5xxhb\" (UID: \"f09087c5-87d5-4cd1-9a91-cbe48713d9e9\") " pod="openstack/horizon-86d9b546cb-5xxhb" Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.689370 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f09087c5-87d5-4cd1-9a91-cbe48713d9e9-combined-ca-bundle\") pod \"horizon-86d9b546cb-5xxhb\" (UID: \"f09087c5-87d5-4cd1-9a91-cbe48713d9e9\") " pod="openstack/horizon-86d9b546cb-5xxhb" Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.693680 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/f09087c5-87d5-4cd1-9a91-cbe48713d9e9-horizon-tls-certs\") pod \"horizon-86d9b546cb-5xxhb\" (UID: \"f09087c5-87d5-4cd1-9a91-cbe48713d9e9\") " pod="openstack/horizon-86d9b546cb-5xxhb" Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.705059 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4b5g\" (UniqueName: \"kubernetes.io/projected/f09087c5-87d5-4cd1-9a91-cbe48713d9e9-kube-api-access-p4b5g\") pod \"horizon-86d9b546cb-5xxhb\" (UID: \"f09087c5-87d5-4cd1-9a91-cbe48713d9e9\") " pod="openstack/horizon-86d9b546cb-5xxhb" Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.755095 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93fb8bea-4246-40da-b480-c980b628e2b9-combined-ca-bundle\") pod \"horizon-7c5488f496-jvw7p\" (UID: \"93fb8bea-4246-40da-b480-c980b628e2b9\") " pod="openstack/horizon-7c5488f496-jvw7p" Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.755343 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/93fb8bea-4246-40da-b480-c980b628e2b9-scripts\") pod \"horizon-7c5488f496-jvw7p\" (UID: \"93fb8bea-4246-40da-b480-c980b628e2b9\") " pod="openstack/horizon-7c5488f496-jvw7p" Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.755633 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/93fb8bea-4246-40da-b480-c980b628e2b9-horizon-tls-certs\") pod \"horizon-7c5488f496-jvw7p\" (UID: \"93fb8bea-4246-40da-b480-c980b628e2b9\") " pod="openstack/horizon-7c5488f496-jvw7p" Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.755669 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/93fb8bea-4246-40da-b480-c980b628e2b9-horizon-secret-key\") pod \"horizon-7c5488f496-jvw7p\" (UID: \"93fb8bea-4246-40da-b480-c980b628e2b9\") " pod="openstack/horizon-7c5488f496-jvw7p" Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.755817 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/93fb8bea-4246-40da-b480-c980b628e2b9-logs\") pod \"horizon-7c5488f496-jvw7p\" (UID: \"93fb8bea-4246-40da-b480-c980b628e2b9\") " pod="openstack/horizon-7c5488f496-jvw7p" Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.755845 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5rmc\" (UniqueName: \"kubernetes.io/projected/93fb8bea-4246-40da-b480-c980b628e2b9-kube-api-access-p5rmc\") pod \"horizon-7c5488f496-jvw7p\" (UID: \"93fb8bea-4246-40da-b480-c980b628e2b9\") " pod="openstack/horizon-7c5488f496-jvw7p" Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.755991 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/93fb8bea-4246-40da-b480-c980b628e2b9-config-data\") pod \"horizon-7c5488f496-jvw7p\" (UID: \"93fb8bea-4246-40da-b480-c980b628e2b9\") " pod="openstack/horizon-7c5488f496-jvw7p" Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.758632 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/93fb8bea-4246-40da-b480-c980b628e2b9-logs\") pod \"horizon-7c5488f496-jvw7p\" (UID: \"93fb8bea-4246-40da-b480-c980b628e2b9\") " pod="openstack/horizon-7c5488f496-jvw7p" Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.759304 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/93fb8bea-4246-40da-b480-c980b628e2b9-config-data\") pod \"horizon-7c5488f496-jvw7p\" (UID: \"93fb8bea-4246-40da-b480-c980b628e2b9\") " pod="openstack/horizon-7c5488f496-jvw7p" Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.759427 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/93fb8bea-4246-40da-b480-c980b628e2b9-scripts\") pod \"horizon-7c5488f496-jvw7p\" (UID: \"93fb8bea-4246-40da-b480-c980b628e2b9\") " pod="openstack/horizon-7c5488f496-jvw7p" Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.774638 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/93fb8bea-4246-40da-b480-c980b628e2b9-horizon-secret-key\") pod \"horizon-7c5488f496-jvw7p\" (UID: \"93fb8bea-4246-40da-b480-c980b628e2b9\") " pod="openstack/horizon-7c5488f496-jvw7p" Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.777708 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/93fb8bea-4246-40da-b480-c980b628e2b9-horizon-tls-certs\") pod \"horizon-7c5488f496-jvw7p\" (UID: \"93fb8bea-4246-40da-b480-c980b628e2b9\") " pod="openstack/horizon-7c5488f496-jvw7p" Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.779001 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93fb8bea-4246-40da-b480-c980b628e2b9-combined-ca-bundle\") pod \"horizon-7c5488f496-jvw7p\" (UID: \"93fb8bea-4246-40da-b480-c980b628e2b9\") " pod="openstack/horizon-7c5488f496-jvw7p" Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.787940 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5rmc\" (UniqueName: \"kubernetes.io/projected/93fb8bea-4246-40da-b480-c980b628e2b9-kube-api-access-p5rmc\") pod \"horizon-7c5488f496-jvw7p\" (UID: \"93fb8bea-4246-40da-b480-c980b628e2b9\") " pod="openstack/horizon-7c5488f496-jvw7p" Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.797750 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-86d9b546cb-5xxhb" Nov 29 05:28:46 crc kubenswrapper[4799]: I1129 05:28:46.925821 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c5488f496-jvw7p" Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.107684 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-787d-account-create-update-9qmsv" event={"ID":"81148106-ce19-4be8-8afd-0db4518977b0","Type":"ContainerDied","Data":"a084ef8b4e965a0331d8e283f5c16b8744629a6fb2ea9c628a186b6e2669d147"} Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.110004 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a084ef8b4e965a0331d8e283f5c16b8744629a6fb2ea9c628a186b6e2669d147" Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.112192 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"67c0f919-41ec-4c54-a8d4-fea37e8c3efb","Type":"ContainerStarted","Data":"5747ef4c2a7c894296a8d0c3e1566e7b5a5d2042e25535da4a8078b1392b7857"} Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.112521 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="67c0f919-41ec-4c54-a8d4-fea37e8c3efb" containerName="glance-log" containerID="cri-o://d20988a5ceb601b7bf3e1991eca4dad35d9042054d41d5047aabed0d25b30acc" gracePeriod=30 Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.113085 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="67c0f919-41ec-4c54-a8d4-fea37e8c3efb" containerName="glance-httpd" containerID="cri-o://5747ef4c2a7c894296a8d0c3e1566e7b5a5d2042e25535da4a8078b1392b7857" gracePeriod=30 Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.126996 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-r7r7q" event={"ID":"9e368006-30f3-4355-86ca-7ac3ec9cbbe7","Type":"ContainerDied","Data":"444d4532ff7e7dd16652742240dec3a3f6067821fdcc9eefb3b7fbd1f78408ca"} Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.127383 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="444d4532ff7e7dd16652742240dec3a3f6067821fdcc9eefb3b7fbd1f78408ca" Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.146998 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.146972352 podStartE2EDuration="4.146972352s" podCreationTimestamp="2025-11-29 05:28:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 05:28:47.142356973 +0000 UTC m=+3002.785287373" watchObservedRunningTime="2025-11-29 05:28:47.146972352 +0000 UTC m=+3002.789902752" Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.163677 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"0907967b-0bf5-4fe0-9862-ffbd54efde7f","Type":"ContainerStarted","Data":"eed718a9a327020a5ce3ae8385c7878d888ba45e362091196fb268166ee2a445"} Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.170685 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"caa349ef-0de7-46c4-a7cd-fdf43aa6f949","Type":"ContainerStarted","Data":"fc34c2be2e51d9bb4b3dfba3631f227882d6c54611c7384eca1de4fd6081a11a"} Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.182812 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-787d-account-create-update-9qmsv" Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.207608 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-volume1-0" podStartSLOduration=3.7294330970000003 podStartE2EDuration="5.20758673s" podCreationTimestamp="2025-11-29 05:28:42 +0000 UTC" firstStartedPulling="2025-11-29 05:28:43.942735171 +0000 UTC m=+2999.585665571" lastFinishedPulling="2025-11-29 05:28:45.420888804 +0000 UTC m=+3001.063819204" observedRunningTime="2025-11-29 05:28:47.193336999 +0000 UTC m=+3002.836267399" watchObservedRunningTime="2025-11-29 05:28:47.20758673 +0000 UTC m=+3002.850517130" Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.246936 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-r7r7q" Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.271521 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-88d65\" (UniqueName: \"kubernetes.io/projected/81148106-ce19-4be8-8afd-0db4518977b0-kube-api-access-88d65\") pod \"81148106-ce19-4be8-8afd-0db4518977b0\" (UID: \"81148106-ce19-4be8-8afd-0db4518977b0\") " Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.271595 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9e368006-30f3-4355-86ca-7ac3ec9cbbe7-operator-scripts\") pod \"9e368006-30f3-4355-86ca-7ac3ec9cbbe7\" (UID: \"9e368006-30f3-4355-86ca-7ac3ec9cbbe7\") " Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.271635 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81148106-ce19-4be8-8afd-0db4518977b0-operator-scripts\") pod \"81148106-ce19-4be8-8afd-0db4518977b0\" (UID: \"81148106-ce19-4be8-8afd-0db4518977b0\") " Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.271724 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vk8wt\" (UniqueName: \"kubernetes.io/projected/9e368006-30f3-4355-86ca-7ac3ec9cbbe7-kube-api-access-vk8wt\") pod \"9e368006-30f3-4355-86ca-7ac3ec9cbbe7\" (UID: \"9e368006-30f3-4355-86ca-7ac3ec9cbbe7\") " Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.274460 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e368006-30f3-4355-86ca-7ac3ec9cbbe7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9e368006-30f3-4355-86ca-7ac3ec9cbbe7" (UID: "9e368006-30f3-4355-86ca-7ac3ec9cbbe7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.276312 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81148106-ce19-4be8-8afd-0db4518977b0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "81148106-ce19-4be8-8afd-0db4518977b0" (UID: "81148106-ce19-4be8-8afd-0db4518977b0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.280498 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e368006-30f3-4355-86ca-7ac3ec9cbbe7-kube-api-access-vk8wt" (OuterVolumeSpecName: "kube-api-access-vk8wt") pod "9e368006-30f3-4355-86ca-7ac3ec9cbbe7" (UID: "9e368006-30f3-4355-86ca-7ac3ec9cbbe7"). InnerVolumeSpecName "kube-api-access-vk8wt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.285097 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81148106-ce19-4be8-8afd-0db4518977b0-kube-api-access-88d65" (OuterVolumeSpecName: "kube-api-access-88d65") pod "81148106-ce19-4be8-8afd-0db4518977b0" (UID: "81148106-ce19-4be8-8afd-0db4518977b0"). InnerVolumeSpecName "kube-api-access-88d65". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.385580 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9e368006-30f3-4355-86ca-7ac3ec9cbbe7-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.386409 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81148106-ce19-4be8-8afd-0db4518977b0-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.386422 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vk8wt\" (UniqueName: \"kubernetes.io/projected/9e368006-30f3-4355-86ca-7ac3ec9cbbe7-kube-api-access-vk8wt\") on node \"crc\" DevicePath \"\"" Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.386621 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-88d65\" (UniqueName: \"kubernetes.io/projected/81148106-ce19-4be8-8afd-0db4518977b0-kube-api-access-88d65\") on node \"crc\" DevicePath \"\"" Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.481122 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-86d9b546cb-5xxhb"] Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.490701 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7c5488f496-jvw7p"] Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.746052 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.795621 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"67c0f919-41ec-4c54-a8d4-fea37e8c3efb\" (UID: \"67c0f919-41ec-4c54-a8d4-fea37e8c3efb\") " Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.795748 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/67c0f919-41ec-4c54-a8d4-fea37e8c3efb-public-tls-certs\") pod \"67c0f919-41ec-4c54-a8d4-fea37e8c3efb\" (UID: \"67c0f919-41ec-4c54-a8d4-fea37e8c3efb\") " Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.796781 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/67c0f919-41ec-4c54-a8d4-fea37e8c3efb-ceph\") pod \"67c0f919-41ec-4c54-a8d4-fea37e8c3efb\" (UID: \"67c0f919-41ec-4c54-a8d4-fea37e8c3efb\") " Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.796866 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dfqfv\" (UniqueName: \"kubernetes.io/projected/67c0f919-41ec-4c54-a8d4-fea37e8c3efb-kube-api-access-dfqfv\") pod \"67c0f919-41ec-4c54-a8d4-fea37e8c3efb\" (UID: \"67c0f919-41ec-4c54-a8d4-fea37e8c3efb\") " Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.797014 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67c0f919-41ec-4c54-a8d4-fea37e8c3efb-config-data\") pod \"67c0f919-41ec-4c54-a8d4-fea37e8c3efb\" (UID: \"67c0f919-41ec-4c54-a8d4-fea37e8c3efb\") " Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.797063 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67c0f919-41ec-4c54-a8d4-fea37e8c3efb-scripts\") pod \"67c0f919-41ec-4c54-a8d4-fea37e8c3efb\" (UID: \"67c0f919-41ec-4c54-a8d4-fea37e8c3efb\") " Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.797109 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67c0f919-41ec-4c54-a8d4-fea37e8c3efb-logs\") pod \"67c0f919-41ec-4c54-a8d4-fea37e8c3efb\" (UID: \"67c0f919-41ec-4c54-a8d4-fea37e8c3efb\") " Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.797136 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/67c0f919-41ec-4c54-a8d4-fea37e8c3efb-httpd-run\") pod \"67c0f919-41ec-4c54-a8d4-fea37e8c3efb\" (UID: \"67c0f919-41ec-4c54-a8d4-fea37e8c3efb\") " Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.797172 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67c0f919-41ec-4c54-a8d4-fea37e8c3efb-combined-ca-bundle\") pod \"67c0f919-41ec-4c54-a8d4-fea37e8c3efb\" (UID: \"67c0f919-41ec-4c54-a8d4-fea37e8c3efb\") " Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.799221 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67c0f919-41ec-4c54-a8d4-fea37e8c3efb-logs" (OuterVolumeSpecName: "logs") pod "67c0f919-41ec-4c54-a8d4-fea37e8c3efb" (UID: "67c0f919-41ec-4c54-a8d4-fea37e8c3efb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.799321 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67c0f919-41ec-4c54-a8d4-fea37e8c3efb-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "67c0f919-41ec-4c54-a8d4-fea37e8c3efb" (UID: "67c0f919-41ec-4c54-a8d4-fea37e8c3efb"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.807910 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67c0f919-41ec-4c54-a8d4-fea37e8c3efb-scripts" (OuterVolumeSpecName: "scripts") pod "67c0f919-41ec-4c54-a8d4-fea37e8c3efb" (UID: "67c0f919-41ec-4c54-a8d4-fea37e8c3efb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.811111 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "67c0f919-41ec-4c54-a8d4-fea37e8c3efb" (UID: "67c0f919-41ec-4c54-a8d4-fea37e8c3efb"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.814846 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67c0f919-41ec-4c54-a8d4-fea37e8c3efb-ceph" (OuterVolumeSpecName: "ceph") pod "67c0f919-41ec-4c54-a8d4-fea37e8c3efb" (UID: "67c0f919-41ec-4c54-a8d4-fea37e8c3efb"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.814996 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67c0f919-41ec-4c54-a8d4-fea37e8c3efb-kube-api-access-dfqfv" (OuterVolumeSpecName: "kube-api-access-dfqfv") pod "67c0f919-41ec-4c54-a8d4-fea37e8c3efb" (UID: "67c0f919-41ec-4c54-a8d4-fea37e8c3efb"). InnerVolumeSpecName "kube-api-access-dfqfv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.864117 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67c0f919-41ec-4c54-a8d4-fea37e8c3efb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "67c0f919-41ec-4c54-a8d4-fea37e8c3efb" (UID: "67c0f919-41ec-4c54-a8d4-fea37e8c3efb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.887027 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67c0f919-41ec-4c54-a8d4-fea37e8c3efb-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "67c0f919-41ec-4c54-a8d4-fea37e8c3efb" (UID: "67c0f919-41ec-4c54-a8d4-fea37e8c3efb"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.899558 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67c0f919-41ec-4c54-a8d4-fea37e8c3efb-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.899584 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67c0f919-41ec-4c54-a8d4-fea37e8c3efb-logs\") on node \"crc\" DevicePath \"\"" Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.899594 4799 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/67c0f919-41ec-4c54-a8d4-fea37e8c3efb-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.899604 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67c0f919-41ec-4c54-a8d4-fea37e8c3efb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.899639 4799 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.899650 4799 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/67c0f919-41ec-4c54-a8d4-fea37e8c3efb-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.899662 4799 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/67c0f919-41ec-4c54-a8d4-fea37e8c3efb-ceph\") on node \"crc\" DevicePath \"\"" Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.899673 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dfqfv\" (UniqueName: \"kubernetes.io/projected/67c0f919-41ec-4c54-a8d4-fea37e8c3efb-kube-api-access-dfqfv\") on node \"crc\" DevicePath \"\"" Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.925082 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67c0f919-41ec-4c54-a8d4-fea37e8c3efb-config-data" (OuterVolumeSpecName: "config-data") pod "67c0f919-41ec-4c54-a8d4-fea37e8c3efb" (UID: "67c0f919-41ec-4c54-a8d4-fea37e8c3efb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.942747 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:47 crc kubenswrapper[4799]: I1129 05:28:47.950735 4799 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.004052 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67c0f919-41ec-4c54-a8d4-fea37e8c3efb-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.004582 4799 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.192326 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="304c844f-62a9-4253-b68b-a5003341cea1" containerName="glance-log" containerID="cri-o://aaf55259341ee3c336f469d5eae5425cff1a95608ba68c660aa005ba5ed84d2d" gracePeriod=30 Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.192841 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="304c844f-62a9-4253-b68b-a5003341cea1" containerName="glance-httpd" containerID="cri-o://9e06364a85a1818eb02dd24e67727d336e3d1be556cd2cffdd3c72caeb100ab5" gracePeriod=30 Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.193143 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"304c844f-62a9-4253-b68b-a5003341cea1","Type":"ContainerStarted","Data":"9e06364a85a1818eb02dd24e67727d336e3d1be556cd2cffdd3c72caeb100ab5"} Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.204578 4799 generic.go:334] "Generic (PLEG): container finished" podID="67c0f919-41ec-4c54-a8d4-fea37e8c3efb" containerID="5747ef4c2a7c894296a8d0c3e1566e7b5a5d2042e25535da4a8078b1392b7857" exitCode=143 Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.204631 4799 generic.go:334] "Generic (PLEG): container finished" podID="67c0f919-41ec-4c54-a8d4-fea37e8c3efb" containerID="d20988a5ceb601b7bf3e1991eca4dad35d9042054d41d5047aabed0d25b30acc" exitCode=143 Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.204715 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"67c0f919-41ec-4c54-a8d4-fea37e8c3efb","Type":"ContainerDied","Data":"5747ef4c2a7c894296a8d0c3e1566e7b5a5d2042e25535da4a8078b1392b7857"} Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.204731 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.204759 4799 scope.go:117] "RemoveContainer" containerID="5747ef4c2a7c894296a8d0c3e1566e7b5a5d2042e25535da4a8078b1392b7857" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.204745 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"67c0f919-41ec-4c54-a8d4-fea37e8c3efb","Type":"ContainerDied","Data":"d20988a5ceb601b7bf3e1991eca4dad35d9042054d41d5047aabed0d25b30acc"} Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.204868 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"67c0f919-41ec-4c54-a8d4-fea37e8c3efb","Type":"ContainerDied","Data":"b90cad992bc106407343de0b2c52191e30df994ae3c54a07fed012509b688d95"} Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.226846 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-86d9b546cb-5xxhb" event={"ID":"f09087c5-87d5-4cd1-9a91-cbe48713d9e9","Type":"ContainerStarted","Data":"c930f52404aeb6dd443f7a9afd2c5b13d6e72dac127362739964a73f6fff9441"} Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.231450 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"caa349ef-0de7-46c4-a7cd-fdf43aa6f949","Type":"ContainerStarted","Data":"35726819f19cbaaec0dea14498cc4303bb903f54ac84fb6859204e2dce765cb8"} Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.233508 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.233396327 podStartE2EDuration="5.233396327s" podCreationTimestamp="2025-11-29 05:28:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 05:28:48.215585473 +0000 UTC m=+3003.858515873" watchObservedRunningTime="2025-11-29 05:28:48.233396327 +0000 UTC m=+3003.876326727" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.239285 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-r7r7q" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.240572 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-787d-account-create-update-9qmsv" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.240282 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c5488f496-jvw7p" event={"ID":"93fb8bea-4246-40da-b480-c980b628e2b9","Type":"ContainerStarted","Data":"a4ade76d74d8165a047f1ff10eba6e12e0caa1096361b9e1cad41f869a808f76"} Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.278145 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=4.665719419 podStartE2EDuration="6.278126305s" podCreationTimestamp="2025-11-29 05:28:42 +0000 UTC" firstStartedPulling="2025-11-29 05:28:43.808298963 +0000 UTC m=+2999.451229363" lastFinishedPulling="2025-11-29 05:28:45.420705849 +0000 UTC m=+3001.063636249" observedRunningTime="2025-11-29 05:28:48.270291989 +0000 UTC m=+3003.913222389" watchObservedRunningTime="2025-11-29 05:28:48.278126305 +0000 UTC m=+3003.921056695" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.282519 4799 scope.go:117] "RemoveContainer" containerID="d20988a5ceb601b7bf3e1991eca4dad35d9042054d41d5047aabed0d25b30acc" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.307540 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.329321 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.353688 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 29 05:28:48 crc kubenswrapper[4799]: E1129 05:28:48.354379 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e368006-30f3-4355-86ca-7ac3ec9cbbe7" containerName="mariadb-database-create" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.354397 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e368006-30f3-4355-86ca-7ac3ec9cbbe7" containerName="mariadb-database-create" Nov 29 05:28:48 crc kubenswrapper[4799]: E1129 05:28:48.354418 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67c0f919-41ec-4c54-a8d4-fea37e8c3efb" containerName="glance-log" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.354424 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="67c0f919-41ec-4c54-a8d4-fea37e8c3efb" containerName="glance-log" Nov 29 05:28:48 crc kubenswrapper[4799]: E1129 05:28:48.354438 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81148106-ce19-4be8-8afd-0db4518977b0" containerName="mariadb-account-create-update" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.354444 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="81148106-ce19-4be8-8afd-0db4518977b0" containerName="mariadb-account-create-update" Nov 29 05:28:48 crc kubenswrapper[4799]: E1129 05:28:48.354467 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67c0f919-41ec-4c54-a8d4-fea37e8c3efb" containerName="glance-httpd" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.354473 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="67c0f919-41ec-4c54-a8d4-fea37e8c3efb" containerName="glance-httpd" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.355873 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="81148106-ce19-4be8-8afd-0db4518977b0" containerName="mariadb-account-create-update" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.355899 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="67c0f919-41ec-4c54-a8d4-fea37e8c3efb" containerName="glance-httpd" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.355910 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="67c0f919-41ec-4c54-a8d4-fea37e8c3efb" containerName="glance-log" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.355930 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e368006-30f3-4355-86ca-7ac3ec9cbbe7" containerName="mariadb-database-create" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.357586 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.362209 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.364058 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.362897 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.377667 4799 scope.go:117] "RemoveContainer" containerID="5747ef4c2a7c894296a8d0c3e1566e7b5a5d2042e25535da4a8078b1392b7857" Nov 29 05:28:48 crc kubenswrapper[4799]: E1129 05:28:48.378398 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5747ef4c2a7c894296a8d0c3e1566e7b5a5d2042e25535da4a8078b1392b7857\": container with ID starting with 5747ef4c2a7c894296a8d0c3e1566e7b5a5d2042e25535da4a8078b1392b7857 not found: ID does not exist" containerID="5747ef4c2a7c894296a8d0c3e1566e7b5a5d2042e25535da4a8078b1392b7857" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.378465 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5747ef4c2a7c894296a8d0c3e1566e7b5a5d2042e25535da4a8078b1392b7857"} err="failed to get container status \"5747ef4c2a7c894296a8d0c3e1566e7b5a5d2042e25535da4a8078b1392b7857\": rpc error: code = NotFound desc = could not find container \"5747ef4c2a7c894296a8d0c3e1566e7b5a5d2042e25535da4a8078b1392b7857\": container with ID starting with 5747ef4c2a7c894296a8d0c3e1566e7b5a5d2042e25535da4a8078b1392b7857 not found: ID does not exist" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.378505 4799 scope.go:117] "RemoveContainer" containerID="d20988a5ceb601b7bf3e1991eca4dad35d9042054d41d5047aabed0d25b30acc" Nov 29 05:28:48 crc kubenswrapper[4799]: E1129 05:28:48.379066 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d20988a5ceb601b7bf3e1991eca4dad35d9042054d41d5047aabed0d25b30acc\": container with ID starting with d20988a5ceb601b7bf3e1991eca4dad35d9042054d41d5047aabed0d25b30acc not found: ID does not exist" containerID="d20988a5ceb601b7bf3e1991eca4dad35d9042054d41d5047aabed0d25b30acc" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.379090 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d20988a5ceb601b7bf3e1991eca4dad35d9042054d41d5047aabed0d25b30acc"} err="failed to get container status \"d20988a5ceb601b7bf3e1991eca4dad35d9042054d41d5047aabed0d25b30acc\": rpc error: code = NotFound desc = could not find container \"d20988a5ceb601b7bf3e1991eca4dad35d9042054d41d5047aabed0d25b30acc\": container with ID starting with d20988a5ceb601b7bf3e1991eca4dad35d9042054d41d5047aabed0d25b30acc not found: ID does not exist" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.379112 4799 scope.go:117] "RemoveContainer" containerID="5747ef4c2a7c894296a8d0c3e1566e7b5a5d2042e25535da4a8078b1392b7857" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.383363 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5747ef4c2a7c894296a8d0c3e1566e7b5a5d2042e25535da4a8078b1392b7857"} err="failed to get container status \"5747ef4c2a7c894296a8d0c3e1566e7b5a5d2042e25535da4a8078b1392b7857\": rpc error: code = NotFound desc = could not find container \"5747ef4c2a7c894296a8d0c3e1566e7b5a5d2042e25535da4a8078b1392b7857\": container with ID starting with 5747ef4c2a7c894296a8d0c3e1566e7b5a5d2042e25535da4a8078b1392b7857 not found: ID does not exist" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.383410 4799 scope.go:117] "RemoveContainer" containerID="d20988a5ceb601b7bf3e1991eca4dad35d9042054d41d5047aabed0d25b30acc" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.384944 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d20988a5ceb601b7bf3e1991eca4dad35d9042054d41d5047aabed0d25b30acc"} err="failed to get container status \"d20988a5ceb601b7bf3e1991eca4dad35d9042054d41d5047aabed0d25b30acc\": rpc error: code = NotFound desc = could not find container \"d20988a5ceb601b7bf3e1991eca4dad35d9042054d41d5047aabed0d25b30acc\": container with ID starting with d20988a5ceb601b7bf3e1991eca4dad35d9042054d41d5047aabed0d25b30acc not found: ID does not exist" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.517937 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd7d046d-4862-47c2-b4d3-e4d81d79706a-scripts\") pod \"glance-default-external-api-0\" (UID: \"cd7d046d-4862-47c2-b4d3-e4d81d79706a\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.518021 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd7d046d-4862-47c2-b4d3-e4d81d79706a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"cd7d046d-4862-47c2-b4d3-e4d81d79706a\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.518085 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cl4s\" (UniqueName: \"kubernetes.io/projected/cd7d046d-4862-47c2-b4d3-e4d81d79706a-kube-api-access-8cl4s\") pod \"glance-default-external-api-0\" (UID: \"cd7d046d-4862-47c2-b4d3-e4d81d79706a\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.518143 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/cd7d046d-4862-47c2-b4d3-e4d81d79706a-ceph\") pod \"glance-default-external-api-0\" (UID: \"cd7d046d-4862-47c2-b4d3-e4d81d79706a\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.518190 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd7d046d-4862-47c2-b4d3-e4d81d79706a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"cd7d046d-4862-47c2-b4d3-e4d81d79706a\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.518243 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"cd7d046d-4862-47c2-b4d3-e4d81d79706a\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.518322 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cd7d046d-4862-47c2-b4d3-e4d81d79706a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"cd7d046d-4862-47c2-b4d3-e4d81d79706a\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.518382 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd7d046d-4862-47c2-b4d3-e4d81d79706a-logs\") pod \"glance-default-external-api-0\" (UID: \"cd7d046d-4862-47c2-b4d3-e4d81d79706a\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.518660 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd7d046d-4862-47c2-b4d3-e4d81d79706a-config-data\") pod \"glance-default-external-api-0\" (UID: \"cd7d046d-4862-47c2-b4d3-e4d81d79706a\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.621505 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd7d046d-4862-47c2-b4d3-e4d81d79706a-logs\") pod \"glance-default-external-api-0\" (UID: \"cd7d046d-4862-47c2-b4d3-e4d81d79706a\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.621566 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd7d046d-4862-47c2-b4d3-e4d81d79706a-config-data\") pod \"glance-default-external-api-0\" (UID: \"cd7d046d-4862-47c2-b4d3-e4d81d79706a\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.621622 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd7d046d-4862-47c2-b4d3-e4d81d79706a-scripts\") pod \"glance-default-external-api-0\" (UID: \"cd7d046d-4862-47c2-b4d3-e4d81d79706a\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.621649 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd7d046d-4862-47c2-b4d3-e4d81d79706a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"cd7d046d-4862-47c2-b4d3-e4d81d79706a\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.621697 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cl4s\" (UniqueName: \"kubernetes.io/projected/cd7d046d-4862-47c2-b4d3-e4d81d79706a-kube-api-access-8cl4s\") pod \"glance-default-external-api-0\" (UID: \"cd7d046d-4862-47c2-b4d3-e4d81d79706a\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.621723 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/cd7d046d-4862-47c2-b4d3-e4d81d79706a-ceph\") pod \"glance-default-external-api-0\" (UID: \"cd7d046d-4862-47c2-b4d3-e4d81d79706a\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.621749 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd7d046d-4862-47c2-b4d3-e4d81d79706a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"cd7d046d-4862-47c2-b4d3-e4d81d79706a\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.621805 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"cd7d046d-4862-47c2-b4d3-e4d81d79706a\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.621942 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cd7d046d-4862-47c2-b4d3-e4d81d79706a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"cd7d046d-4862-47c2-b4d3-e4d81d79706a\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.622258 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cd7d046d-4862-47c2-b4d3-e4d81d79706a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"cd7d046d-4862-47c2-b4d3-e4d81d79706a\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.622611 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"cd7d046d-4862-47c2-b4d3-e4d81d79706a\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.623109 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd7d046d-4862-47c2-b4d3-e4d81d79706a-logs\") pod \"glance-default-external-api-0\" (UID: \"cd7d046d-4862-47c2-b4d3-e4d81d79706a\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.634062 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/cd7d046d-4862-47c2-b4d3-e4d81d79706a-ceph\") pod \"glance-default-external-api-0\" (UID: \"cd7d046d-4862-47c2-b4d3-e4d81d79706a\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.637061 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd7d046d-4862-47c2-b4d3-e4d81d79706a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"cd7d046d-4862-47c2-b4d3-e4d81d79706a\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.637354 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd7d046d-4862-47c2-b4d3-e4d81d79706a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"cd7d046d-4862-47c2-b4d3-e4d81d79706a\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.640594 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd7d046d-4862-47c2-b4d3-e4d81d79706a-scripts\") pod \"glance-default-external-api-0\" (UID: \"cd7d046d-4862-47c2-b4d3-e4d81d79706a\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.654390 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd7d046d-4862-47c2-b4d3-e4d81d79706a-config-data\") pod \"glance-default-external-api-0\" (UID: \"cd7d046d-4862-47c2-b4d3-e4d81d79706a\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.655843 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cl4s\" (UniqueName: \"kubernetes.io/projected/cd7d046d-4862-47c2-b4d3-e4d81d79706a-kube-api-access-8cl4s\") pod \"glance-default-external-api-0\" (UID: \"cd7d046d-4862-47c2-b4d3-e4d81d79706a\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.685047 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"cd7d046d-4862-47c2-b4d3-e4d81d79706a\") " pod="openstack/glance-default-external-api-0" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.726530 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67c0f919-41ec-4c54-a8d4-fea37e8c3efb" path="/var/lib/kubelet/pods/67c0f919-41ec-4c54-a8d4-fea37e8c3efb/volumes" Nov 29 05:28:48 crc kubenswrapper[4799]: I1129 05:28:48.754222 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.108600 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.261422 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/304c844f-62a9-4253-b68b-a5003341cea1-ceph\") pod \"304c844f-62a9-4253-b68b-a5003341cea1\" (UID: \"304c844f-62a9-4253-b68b-a5003341cea1\") " Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.262225 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/304c844f-62a9-4253-b68b-a5003341cea1-httpd-run\") pod \"304c844f-62a9-4253-b68b-a5003341cea1\" (UID: \"304c844f-62a9-4253-b68b-a5003341cea1\") " Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.262281 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/304c844f-62a9-4253-b68b-a5003341cea1-scripts\") pod \"304c844f-62a9-4253-b68b-a5003341cea1\" (UID: \"304c844f-62a9-4253-b68b-a5003341cea1\") " Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.262304 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"304c844f-62a9-4253-b68b-a5003341cea1\" (UID: \"304c844f-62a9-4253-b68b-a5003341cea1\") " Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.262330 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jdss\" (UniqueName: \"kubernetes.io/projected/304c844f-62a9-4253-b68b-a5003341cea1-kube-api-access-8jdss\") pod \"304c844f-62a9-4253-b68b-a5003341cea1\" (UID: \"304c844f-62a9-4253-b68b-a5003341cea1\") " Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.262365 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/304c844f-62a9-4253-b68b-a5003341cea1-logs\") pod \"304c844f-62a9-4253-b68b-a5003341cea1\" (UID: \"304c844f-62a9-4253-b68b-a5003341cea1\") " Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.262397 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/304c844f-62a9-4253-b68b-a5003341cea1-internal-tls-certs\") pod \"304c844f-62a9-4253-b68b-a5003341cea1\" (UID: \"304c844f-62a9-4253-b68b-a5003341cea1\") " Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.262509 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/304c844f-62a9-4253-b68b-a5003341cea1-combined-ca-bundle\") pod \"304c844f-62a9-4253-b68b-a5003341cea1\" (UID: \"304c844f-62a9-4253-b68b-a5003341cea1\") " Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.262533 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/304c844f-62a9-4253-b68b-a5003341cea1-config-data\") pod \"304c844f-62a9-4253-b68b-a5003341cea1\" (UID: \"304c844f-62a9-4253-b68b-a5003341cea1\") " Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.263941 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/304c844f-62a9-4253-b68b-a5003341cea1-logs" (OuterVolumeSpecName: "logs") pod "304c844f-62a9-4253-b68b-a5003341cea1" (UID: "304c844f-62a9-4253-b68b-a5003341cea1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.270016 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/304c844f-62a9-4253-b68b-a5003341cea1-ceph" (OuterVolumeSpecName: "ceph") pod "304c844f-62a9-4253-b68b-a5003341cea1" (UID: "304c844f-62a9-4253-b68b-a5003341cea1"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.270027 4799 generic.go:334] "Generic (PLEG): container finished" podID="304c844f-62a9-4253-b68b-a5003341cea1" containerID="9e06364a85a1818eb02dd24e67727d336e3d1be556cd2cffdd3c72caeb100ab5" exitCode=0 Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.270100 4799 generic.go:334] "Generic (PLEG): container finished" podID="304c844f-62a9-4253-b68b-a5003341cea1" containerID="aaf55259341ee3c336f469d5eae5425cff1a95608ba68c660aa005ba5ed84d2d" exitCode=143 Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.270141 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.270226 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"304c844f-62a9-4253-b68b-a5003341cea1","Type":"ContainerDied","Data":"9e06364a85a1818eb02dd24e67727d336e3d1be556cd2cffdd3c72caeb100ab5"} Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.270270 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"304c844f-62a9-4253-b68b-a5003341cea1","Type":"ContainerDied","Data":"aaf55259341ee3c336f469d5eae5425cff1a95608ba68c660aa005ba5ed84d2d"} Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.270283 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"304c844f-62a9-4253-b68b-a5003341cea1","Type":"ContainerDied","Data":"96247213ded49154e66366af087376955e1b6cae935b311ca1bd731d50477a4b"} Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.270303 4799 scope.go:117] "RemoveContainer" containerID="9e06364a85a1818eb02dd24e67727d336e3d1be556cd2cffdd3c72caeb100ab5" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.270443 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/304c844f-62a9-4253-b68b-a5003341cea1-kube-api-access-8jdss" (OuterVolumeSpecName: "kube-api-access-8jdss") pod "304c844f-62a9-4253-b68b-a5003341cea1" (UID: "304c844f-62a9-4253-b68b-a5003341cea1"). InnerVolumeSpecName "kube-api-access-8jdss". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.272597 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/304c844f-62a9-4253-b68b-a5003341cea1-scripts" (OuterVolumeSpecName: "scripts") pod "304c844f-62a9-4253-b68b-a5003341cea1" (UID: "304c844f-62a9-4253-b68b-a5003341cea1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.274455 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/304c844f-62a9-4253-b68b-a5003341cea1-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "304c844f-62a9-4253-b68b-a5003341cea1" (UID: "304c844f-62a9-4253-b68b-a5003341cea1"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.280738 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "304c844f-62a9-4253-b68b-a5003341cea1" (UID: "304c844f-62a9-4253-b68b-a5003341cea1"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.315907 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/304c844f-62a9-4253-b68b-a5003341cea1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "304c844f-62a9-4253-b68b-a5003341cea1" (UID: "304c844f-62a9-4253-b68b-a5003341cea1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.342065 4799 scope.go:117] "RemoveContainer" containerID="aaf55259341ee3c336f469d5eae5425cff1a95608ba68c660aa005ba5ed84d2d" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.361048 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/304c844f-62a9-4253-b68b-a5003341cea1-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "304c844f-62a9-4253-b68b-a5003341cea1" (UID: "304c844f-62a9-4253-b68b-a5003341cea1"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.364818 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/304c844f-62a9-4253-b68b-a5003341cea1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.364858 4799 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/304c844f-62a9-4253-b68b-a5003341cea1-ceph\") on node \"crc\" DevicePath \"\"" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.364867 4799 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/304c844f-62a9-4253-b68b-a5003341cea1-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.364878 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/304c844f-62a9-4253-b68b-a5003341cea1-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.364901 4799 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.364914 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jdss\" (UniqueName: \"kubernetes.io/projected/304c844f-62a9-4253-b68b-a5003341cea1-kube-api-access-8jdss\") on node \"crc\" DevicePath \"\"" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.364924 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/304c844f-62a9-4253-b68b-a5003341cea1-logs\") on node \"crc\" DevicePath \"\"" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.364933 4799 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/304c844f-62a9-4253-b68b-a5003341cea1-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.372064 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/304c844f-62a9-4253-b68b-a5003341cea1-config-data" (OuterVolumeSpecName: "config-data") pod "304c844f-62a9-4253-b68b-a5003341cea1" (UID: "304c844f-62a9-4253-b68b-a5003341cea1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.377775 4799 scope.go:117] "RemoveContainer" containerID="9e06364a85a1818eb02dd24e67727d336e3d1be556cd2cffdd3c72caeb100ab5" Nov 29 05:28:49 crc kubenswrapper[4799]: E1129 05:28:49.380297 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e06364a85a1818eb02dd24e67727d336e3d1be556cd2cffdd3c72caeb100ab5\": container with ID starting with 9e06364a85a1818eb02dd24e67727d336e3d1be556cd2cffdd3c72caeb100ab5 not found: ID does not exist" containerID="9e06364a85a1818eb02dd24e67727d336e3d1be556cd2cffdd3c72caeb100ab5" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.380340 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e06364a85a1818eb02dd24e67727d336e3d1be556cd2cffdd3c72caeb100ab5"} err="failed to get container status \"9e06364a85a1818eb02dd24e67727d336e3d1be556cd2cffdd3c72caeb100ab5\": rpc error: code = NotFound desc = could not find container \"9e06364a85a1818eb02dd24e67727d336e3d1be556cd2cffdd3c72caeb100ab5\": container with ID starting with 9e06364a85a1818eb02dd24e67727d336e3d1be556cd2cffdd3c72caeb100ab5 not found: ID does not exist" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.380367 4799 scope.go:117] "RemoveContainer" containerID="aaf55259341ee3c336f469d5eae5425cff1a95608ba68c660aa005ba5ed84d2d" Nov 29 05:28:49 crc kubenswrapper[4799]: E1129 05:28:49.380758 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aaf55259341ee3c336f469d5eae5425cff1a95608ba68c660aa005ba5ed84d2d\": container with ID starting with aaf55259341ee3c336f469d5eae5425cff1a95608ba68c660aa005ba5ed84d2d not found: ID does not exist" containerID="aaf55259341ee3c336f469d5eae5425cff1a95608ba68c660aa005ba5ed84d2d" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.380782 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aaf55259341ee3c336f469d5eae5425cff1a95608ba68c660aa005ba5ed84d2d"} err="failed to get container status \"aaf55259341ee3c336f469d5eae5425cff1a95608ba68c660aa005ba5ed84d2d\": rpc error: code = NotFound desc = could not find container \"aaf55259341ee3c336f469d5eae5425cff1a95608ba68c660aa005ba5ed84d2d\": container with ID starting with aaf55259341ee3c336f469d5eae5425cff1a95608ba68c660aa005ba5ed84d2d not found: ID does not exist" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.380811 4799 scope.go:117] "RemoveContainer" containerID="9e06364a85a1818eb02dd24e67727d336e3d1be556cd2cffdd3c72caeb100ab5" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.381033 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e06364a85a1818eb02dd24e67727d336e3d1be556cd2cffdd3c72caeb100ab5"} err="failed to get container status \"9e06364a85a1818eb02dd24e67727d336e3d1be556cd2cffdd3c72caeb100ab5\": rpc error: code = NotFound desc = could not find container \"9e06364a85a1818eb02dd24e67727d336e3d1be556cd2cffdd3c72caeb100ab5\": container with ID starting with 9e06364a85a1818eb02dd24e67727d336e3d1be556cd2cffdd3c72caeb100ab5 not found: ID does not exist" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.381057 4799 scope.go:117] "RemoveContainer" containerID="aaf55259341ee3c336f469d5eae5425cff1a95608ba68c660aa005ba5ed84d2d" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.381430 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aaf55259341ee3c336f469d5eae5425cff1a95608ba68c660aa005ba5ed84d2d"} err="failed to get container status \"aaf55259341ee3c336f469d5eae5425cff1a95608ba68c660aa005ba5ed84d2d\": rpc error: code = NotFound desc = could not find container \"aaf55259341ee3c336f469d5eae5425cff1a95608ba68c660aa005ba5ed84d2d\": container with ID starting with aaf55259341ee3c336f469d5eae5425cff1a95608ba68c660aa005ba5ed84d2d not found: ID does not exist" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.387138 4799 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.470124 4799 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.470680 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/304c844f-62a9-4253-b68b-a5003341cea1-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.635069 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.663243 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.704481 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 29 05:28:49 crc kubenswrapper[4799]: E1129 05:28:49.705208 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="304c844f-62a9-4253-b68b-a5003341cea1" containerName="glance-log" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.705224 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="304c844f-62a9-4253-b68b-a5003341cea1" containerName="glance-log" Nov 29 05:28:49 crc kubenswrapper[4799]: E1129 05:28:49.705238 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="304c844f-62a9-4253-b68b-a5003341cea1" containerName="glance-httpd" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.705246 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="304c844f-62a9-4253-b68b-a5003341cea1" containerName="glance-httpd" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.705485 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="304c844f-62a9-4253-b68b-a5003341cea1" containerName="glance-log" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.705499 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="304c844f-62a9-4253-b68b-a5003341cea1" containerName="glance-httpd" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.707463 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.711236 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.711607 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.734600 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.805539 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 29 05:28:49 crc kubenswrapper[4799]: W1129 05:28:49.806077 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd7d046d_4862_47c2_b4d3_e4d81d79706a.slice/crio-8a51fac8a7a3450b29a534aeae51e919e96c3bfa7fb5dc26a8c0d641da09820b WatchSource:0}: Error finding container 8a51fac8a7a3450b29a534aeae51e919e96c3bfa7fb5dc26a8c0d641da09820b: Status 404 returned error can't find the container with id 8a51fac8a7a3450b29a534aeae51e919e96c3bfa7fb5dc26a8c0d641da09820b Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.881981 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/477f0d92-5648-405b-8034-f851200268a9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"477f0d92-5648-405b-8034-f851200268a9\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.882035 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/477f0d92-5648-405b-8034-f851200268a9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"477f0d92-5648-405b-8034-f851200268a9\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.882069 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/477f0d92-5648-405b-8034-f851200268a9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"477f0d92-5648-405b-8034-f851200268a9\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.882097 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/477f0d92-5648-405b-8034-f851200268a9-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"477f0d92-5648-405b-8034-f851200268a9\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.882146 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/477f0d92-5648-405b-8034-f851200268a9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"477f0d92-5648-405b-8034-f851200268a9\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.882181 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/477f0d92-5648-405b-8034-f851200268a9-logs\") pod \"glance-default-internal-api-0\" (UID: \"477f0d92-5648-405b-8034-f851200268a9\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.882208 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"477f0d92-5648-405b-8034-f851200268a9\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.882228 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/477f0d92-5648-405b-8034-f851200268a9-ceph\") pod \"glance-default-internal-api-0\" (UID: \"477f0d92-5648-405b-8034-f851200268a9\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.882272 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49bdz\" (UniqueName: \"kubernetes.io/projected/477f0d92-5648-405b-8034-f851200268a9-kube-api-access-49bdz\") pod \"glance-default-internal-api-0\" (UID: \"477f0d92-5648-405b-8034-f851200268a9\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.984613 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/477f0d92-5648-405b-8034-f851200268a9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"477f0d92-5648-405b-8034-f851200268a9\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.984686 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/477f0d92-5648-405b-8034-f851200268a9-logs\") pod \"glance-default-internal-api-0\" (UID: \"477f0d92-5648-405b-8034-f851200268a9\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.984709 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"477f0d92-5648-405b-8034-f851200268a9\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.984734 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/477f0d92-5648-405b-8034-f851200268a9-ceph\") pod \"glance-default-internal-api-0\" (UID: \"477f0d92-5648-405b-8034-f851200268a9\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.984782 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49bdz\" (UniqueName: \"kubernetes.io/projected/477f0d92-5648-405b-8034-f851200268a9-kube-api-access-49bdz\") pod \"glance-default-internal-api-0\" (UID: \"477f0d92-5648-405b-8034-f851200268a9\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.984862 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/477f0d92-5648-405b-8034-f851200268a9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"477f0d92-5648-405b-8034-f851200268a9\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.984927 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/477f0d92-5648-405b-8034-f851200268a9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"477f0d92-5648-405b-8034-f851200268a9\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.984957 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/477f0d92-5648-405b-8034-f851200268a9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"477f0d92-5648-405b-8034-f851200268a9\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.984999 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/477f0d92-5648-405b-8034-f851200268a9-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"477f0d92-5648-405b-8034-f851200268a9\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.985476 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"477f0d92-5648-405b-8034-f851200268a9\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.986469 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/477f0d92-5648-405b-8034-f851200268a9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"477f0d92-5648-405b-8034-f851200268a9\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.987177 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/477f0d92-5648-405b-8034-f851200268a9-logs\") pod \"glance-default-internal-api-0\" (UID: \"477f0d92-5648-405b-8034-f851200268a9\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.994762 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/477f0d92-5648-405b-8034-f851200268a9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"477f0d92-5648-405b-8034-f851200268a9\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.995028 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/477f0d92-5648-405b-8034-f851200268a9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"477f0d92-5648-405b-8034-f851200268a9\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.996470 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/477f0d92-5648-405b-8034-f851200268a9-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"477f0d92-5648-405b-8034-f851200268a9\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:49 crc kubenswrapper[4799]: I1129 05:28:49.998318 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/477f0d92-5648-405b-8034-f851200268a9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"477f0d92-5648-405b-8034-f851200268a9\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:50 crc kubenswrapper[4799]: I1129 05:28:50.010180 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/477f0d92-5648-405b-8034-f851200268a9-ceph\") pod \"glance-default-internal-api-0\" (UID: \"477f0d92-5648-405b-8034-f851200268a9\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:50 crc kubenswrapper[4799]: I1129 05:28:50.039993 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49bdz\" (UniqueName: \"kubernetes.io/projected/477f0d92-5648-405b-8034-f851200268a9-kube-api-access-49bdz\") pod \"glance-default-internal-api-0\" (UID: \"477f0d92-5648-405b-8034-f851200268a9\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:50 crc kubenswrapper[4799]: I1129 05:28:50.087194 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"477f0d92-5648-405b-8034-f851200268a9\") " pod="openstack/glance-default-internal-api-0" Nov 29 05:28:50 crc kubenswrapper[4799]: I1129 05:28:50.294439 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cd7d046d-4862-47c2-b4d3-e4d81d79706a","Type":"ContainerStarted","Data":"8a51fac8a7a3450b29a534aeae51e919e96c3bfa7fb5dc26a8c0d641da09820b"} Nov 29 05:28:50 crc kubenswrapper[4799]: I1129 05:28:50.350860 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 29 05:28:50 crc kubenswrapper[4799]: I1129 05:28:50.671738 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="304c844f-62a9-4253-b68b-a5003341cea1" path="/var/lib/kubelet/pods/304c844f-62a9-4253-b68b-a5003341cea1/volumes" Nov 29 05:28:50 crc kubenswrapper[4799]: I1129 05:28:50.974873 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 29 05:28:50 crc kubenswrapper[4799]: W1129 05:28:50.984777 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod477f0d92_5648_405b_8034_f851200268a9.slice/crio-a068d302074dd9fd441eb619a4d547db633ef77ce1432cb62bffc014b4256089 WatchSource:0}: Error finding container a068d302074dd9fd441eb619a4d547db633ef77ce1432cb62bffc014b4256089: Status 404 returned error can't find the container with id a068d302074dd9fd441eb619a4d547db633ef77ce1432cb62bffc014b4256089 Nov 29 05:28:51 crc kubenswrapper[4799]: I1129 05:28:51.333214 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cd7d046d-4862-47c2-b4d3-e4d81d79706a","Type":"ContainerStarted","Data":"1b94d7ea312daf48e986eccad45018776e145c9297c9d4eb1f33addc19a904d5"} Nov 29 05:28:51 crc kubenswrapper[4799]: I1129 05:28:51.337500 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"477f0d92-5648-405b-8034-f851200268a9","Type":"ContainerStarted","Data":"a068d302074dd9fd441eb619a4d547db633ef77ce1432cb62bffc014b4256089"} Nov 29 05:28:52 crc kubenswrapper[4799]: I1129 05:28:52.354369 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"477f0d92-5648-405b-8034-f851200268a9","Type":"ContainerStarted","Data":"5747be426fca79975325428d544721fd4848ade273ea8d8ac352cb614d27ee84"} Nov 29 05:28:52 crc kubenswrapper[4799]: I1129 05:28:52.361900 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cd7d046d-4862-47c2-b4d3-e4d81d79706a","Type":"ContainerStarted","Data":"37b9744e924d5d92214b7c293b78d04465076283c4f620933fabefd20beb4ec9"} Nov 29 05:28:52 crc kubenswrapper[4799]: I1129 05:28:52.406305 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.406281344 podStartE2EDuration="4.406281344s" podCreationTimestamp="2025-11-29 05:28:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 05:28:52.387893565 +0000 UTC m=+3008.030823965" watchObservedRunningTime="2025-11-29 05:28:52.406281344 +0000 UTC m=+3008.049211744" Nov 29 05:28:52 crc kubenswrapper[4799]: I1129 05:28:52.944991 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Nov 29 05:28:53 crc kubenswrapper[4799]: I1129 05:28:53.255439 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-volume1-0" Nov 29 05:28:53 crc kubenswrapper[4799]: I1129 05:28:53.272432 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Nov 29 05:28:53 crc kubenswrapper[4799]: I1129 05:28:53.553246 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-sync-jlhf6"] Nov 29 05:28:53 crc kubenswrapper[4799]: I1129 05:28:53.555822 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-jlhf6" Nov 29 05:28:53 crc kubenswrapper[4799]: I1129 05:28:53.558731 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Nov 29 05:28:53 crc kubenswrapper[4799]: I1129 05:28:53.559263 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-nngct" Nov 29 05:28:53 crc kubenswrapper[4799]: I1129 05:28:53.576259 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-jlhf6"] Nov 29 05:28:53 crc kubenswrapper[4799]: I1129 05:28:53.611207 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a730664-a7b6-4cbd-b657-d6f8ef3803b8-config-data\") pod \"manila-db-sync-jlhf6\" (UID: \"4a730664-a7b6-4cbd-b657-d6f8ef3803b8\") " pod="openstack/manila-db-sync-jlhf6" Nov 29 05:28:53 crc kubenswrapper[4799]: I1129 05:28:53.611365 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjkj2\" (UniqueName: \"kubernetes.io/projected/4a730664-a7b6-4cbd-b657-d6f8ef3803b8-kube-api-access-bjkj2\") pod \"manila-db-sync-jlhf6\" (UID: \"4a730664-a7b6-4cbd-b657-d6f8ef3803b8\") " pod="openstack/manila-db-sync-jlhf6" Nov 29 05:28:53 crc kubenswrapper[4799]: I1129 05:28:53.611420 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a730664-a7b6-4cbd-b657-d6f8ef3803b8-combined-ca-bundle\") pod \"manila-db-sync-jlhf6\" (UID: \"4a730664-a7b6-4cbd-b657-d6f8ef3803b8\") " pod="openstack/manila-db-sync-jlhf6" Nov 29 05:28:53 crc kubenswrapper[4799]: I1129 05:28:53.611490 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/4a730664-a7b6-4cbd-b657-d6f8ef3803b8-job-config-data\") pod \"manila-db-sync-jlhf6\" (UID: \"4a730664-a7b6-4cbd-b657-d6f8ef3803b8\") " pod="openstack/manila-db-sync-jlhf6" Nov 29 05:28:53 crc kubenswrapper[4799]: I1129 05:28:53.713600 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjkj2\" (UniqueName: \"kubernetes.io/projected/4a730664-a7b6-4cbd-b657-d6f8ef3803b8-kube-api-access-bjkj2\") pod \"manila-db-sync-jlhf6\" (UID: \"4a730664-a7b6-4cbd-b657-d6f8ef3803b8\") " pod="openstack/manila-db-sync-jlhf6" Nov 29 05:28:53 crc kubenswrapper[4799]: I1129 05:28:53.713682 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a730664-a7b6-4cbd-b657-d6f8ef3803b8-combined-ca-bundle\") pod \"manila-db-sync-jlhf6\" (UID: \"4a730664-a7b6-4cbd-b657-d6f8ef3803b8\") " pod="openstack/manila-db-sync-jlhf6" Nov 29 05:28:53 crc kubenswrapper[4799]: I1129 05:28:53.713765 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/4a730664-a7b6-4cbd-b657-d6f8ef3803b8-job-config-data\") pod \"manila-db-sync-jlhf6\" (UID: \"4a730664-a7b6-4cbd-b657-d6f8ef3803b8\") " pod="openstack/manila-db-sync-jlhf6" Nov 29 05:28:53 crc kubenswrapper[4799]: I1129 05:28:53.713824 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a730664-a7b6-4cbd-b657-d6f8ef3803b8-config-data\") pod \"manila-db-sync-jlhf6\" (UID: \"4a730664-a7b6-4cbd-b657-d6f8ef3803b8\") " pod="openstack/manila-db-sync-jlhf6" Nov 29 05:28:53 crc kubenswrapper[4799]: I1129 05:28:53.726058 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/4a730664-a7b6-4cbd-b657-d6f8ef3803b8-job-config-data\") pod \"manila-db-sync-jlhf6\" (UID: \"4a730664-a7b6-4cbd-b657-d6f8ef3803b8\") " pod="openstack/manila-db-sync-jlhf6" Nov 29 05:28:53 crc kubenswrapper[4799]: I1129 05:28:53.728977 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a730664-a7b6-4cbd-b657-d6f8ef3803b8-combined-ca-bundle\") pod \"manila-db-sync-jlhf6\" (UID: \"4a730664-a7b6-4cbd-b657-d6f8ef3803b8\") " pod="openstack/manila-db-sync-jlhf6" Nov 29 05:28:53 crc kubenswrapper[4799]: I1129 05:28:53.729369 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a730664-a7b6-4cbd-b657-d6f8ef3803b8-config-data\") pod \"manila-db-sync-jlhf6\" (UID: \"4a730664-a7b6-4cbd-b657-d6f8ef3803b8\") " pod="openstack/manila-db-sync-jlhf6" Nov 29 05:28:53 crc kubenswrapper[4799]: I1129 05:28:53.731422 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjkj2\" (UniqueName: \"kubernetes.io/projected/4a730664-a7b6-4cbd-b657-d6f8ef3803b8-kube-api-access-bjkj2\") pod \"manila-db-sync-jlhf6\" (UID: \"4a730664-a7b6-4cbd-b657-d6f8ef3803b8\") " pod="openstack/manila-db-sync-jlhf6" Nov 29 05:28:53 crc kubenswrapper[4799]: I1129 05:28:53.886013 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-jlhf6" Nov 29 05:28:54 crc kubenswrapper[4799]: I1129 05:28:54.661481 4799 scope.go:117] "RemoveContainer" containerID="e6d4b3d1d5b5e1eeb13508d67de3d3c3aac52e76337b55e53793185c0d001d82" Nov 29 05:28:54 crc kubenswrapper[4799]: E1129 05:28:54.662324 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:28:58 crc kubenswrapper[4799]: I1129 05:28:58.451176 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-566ff59745-jp249" event={"ID":"ebcbc04a-230e-46b1-be0d-af9ece769369","Type":"ContainerStarted","Data":"a8ed286597384b45acd04d7e67e5c08d350f935fef4e4fbb30c795f46c0acf50"} Nov 29 05:28:58 crc kubenswrapper[4799]: I1129 05:28:58.454747 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-86d9b546cb-5xxhb" event={"ID":"f09087c5-87d5-4cd1-9a91-cbe48713d9e9","Type":"ContainerStarted","Data":"8f1f666f1cb9af792d04255e0b4e5900be38a9d0b4060fd4c97f314b6c12a164"} Nov 29 05:28:58 crc kubenswrapper[4799]: I1129 05:28:58.458241 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-598444b44c-v5znr" event={"ID":"04de4a38-5d09-4c68-b7a2-be6482748e50","Type":"ContainerStarted","Data":"69cac5b4704d7a992b979fd8a84a6d26615f488bf691883048255f969711a221"} Nov 29 05:28:58 crc kubenswrapper[4799]: I1129 05:28:58.460510 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c5488f496-jvw7p" event={"ID":"93fb8bea-4246-40da-b480-c980b628e2b9","Type":"ContainerStarted","Data":"32c917ba032fdf09a306cdbc72e14fc165a2baf80fe79c8ed4c4b76a916406ce"} Nov 29 05:28:58 crc kubenswrapper[4799]: I1129 05:28:58.511901 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-jlhf6"] Nov 29 05:28:58 crc kubenswrapper[4799]: I1129 05:28:58.757064 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 29 05:28:58 crc kubenswrapper[4799]: I1129 05:28:58.757581 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 29 05:28:58 crc kubenswrapper[4799]: I1129 05:28:58.801563 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 29 05:28:58 crc kubenswrapper[4799]: I1129 05:28:58.805684 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 29 05:28:59 crc kubenswrapper[4799]: I1129 05:28:59.481848 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-566ff59745-jp249" podUID="ebcbc04a-230e-46b1-be0d-af9ece769369" containerName="horizon-log" containerID="cri-o://a8ed286597384b45acd04d7e67e5c08d350f935fef4e4fbb30c795f46c0acf50" gracePeriod=30 Nov 29 05:28:59 crc kubenswrapper[4799]: I1129 05:28:59.482275 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-566ff59745-jp249" event={"ID":"ebcbc04a-230e-46b1-be0d-af9ece769369","Type":"ContainerStarted","Data":"72186f35f25c7e3357a6a3815736e28ff75be3b34cc5516c50542e4d500121dc"} Nov 29 05:28:59 crc kubenswrapper[4799]: I1129 05:28:59.482503 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-566ff59745-jp249" podUID="ebcbc04a-230e-46b1-be0d-af9ece769369" containerName="horizon" containerID="cri-o://72186f35f25c7e3357a6a3815736e28ff75be3b34cc5516c50542e4d500121dc" gracePeriod=30 Nov 29 05:28:59 crc kubenswrapper[4799]: I1129 05:28:59.505281 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-jlhf6" event={"ID":"4a730664-a7b6-4cbd-b657-d6f8ef3803b8","Type":"ContainerStarted","Data":"6ccf320d5f0be9ea6360f97ded6dff3de3ee6e7fe45cacc4a93c7a13de90db9d"} Nov 29 05:28:59 crc kubenswrapper[4799]: I1129 05:28:59.518568 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"477f0d92-5648-405b-8034-f851200268a9","Type":"ContainerStarted","Data":"e006a9db25f7bf06704dff86fe217d2ef818ebdc8a52e023f3ec6f043f3cfe04"} Nov 29 05:28:59 crc kubenswrapper[4799]: I1129 05:28:59.523047 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-86d9b546cb-5xxhb" event={"ID":"f09087c5-87d5-4cd1-9a91-cbe48713d9e9","Type":"ContainerStarted","Data":"acf62f8dda4e17622de5c459a6b4fe1c70a55e3481d77ef29c8a3bc868f9df26"} Nov 29 05:28:59 crc kubenswrapper[4799]: I1129 05:28:59.537266 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-598444b44c-v5znr" event={"ID":"04de4a38-5d09-4c68-b7a2-be6482748e50","Type":"ContainerStarted","Data":"f8801fb2760fa321e80c168af0997ee85b677223a66f5991257e83509f823137"} Nov 29 05:28:59 crc kubenswrapper[4799]: I1129 05:28:59.537507 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-598444b44c-v5znr" podUID="04de4a38-5d09-4c68-b7a2-be6482748e50" containerName="horizon-log" containerID="cri-o://69cac5b4704d7a992b979fd8a84a6d26615f488bf691883048255f969711a221" gracePeriod=30 Nov 29 05:28:59 crc kubenswrapper[4799]: I1129 05:28:59.537731 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-598444b44c-v5znr" podUID="04de4a38-5d09-4c68-b7a2-be6482748e50" containerName="horizon" containerID="cri-o://f8801fb2760fa321e80c168af0997ee85b677223a66f5991257e83509f823137" gracePeriod=30 Nov 29 05:28:59 crc kubenswrapper[4799]: I1129 05:28:59.542626 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-566ff59745-jp249" podStartSLOduration=2.633819944 podStartE2EDuration="16.542606685s" podCreationTimestamp="2025-11-29 05:28:43 +0000 UTC" firstStartedPulling="2025-11-29 05:28:44.11492166 +0000 UTC m=+2999.757852060" lastFinishedPulling="2025-11-29 05:28:58.023708401 +0000 UTC m=+3013.666638801" observedRunningTime="2025-11-29 05:28:59.53187382 +0000 UTC m=+3015.174804220" watchObservedRunningTime="2025-11-29 05:28:59.542606685 +0000 UTC m=+3015.185537075" Nov 29 05:28:59 crc kubenswrapper[4799]: I1129 05:28:59.546759 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c5488f496-jvw7p" event={"ID":"93fb8bea-4246-40da-b480-c980b628e2b9","Type":"ContainerStarted","Data":"c396c04a28a8602168622a8170fafd41bc7270a854e8ae4c79b5c207fe83a7d9"} Nov 29 05:28:59 crc kubenswrapper[4799]: I1129 05:28:59.547330 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 29 05:28:59 crc kubenswrapper[4799]: I1129 05:28:59.547380 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 29 05:28:59 crc kubenswrapper[4799]: I1129 05:28:59.566867 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-86d9b546cb-5xxhb" podStartSLOduration=3.040023076 podStartE2EDuration="13.566845405s" podCreationTimestamp="2025-11-29 05:28:46 +0000 UTC" firstStartedPulling="2025-11-29 05:28:47.505442096 +0000 UTC m=+3003.148372496" lastFinishedPulling="2025-11-29 05:28:58.032264425 +0000 UTC m=+3013.675194825" observedRunningTime="2025-11-29 05:28:59.555553065 +0000 UTC m=+3015.198483465" watchObservedRunningTime="2025-11-29 05:28:59.566845405 +0000 UTC m=+3015.209775805" Nov 29 05:28:59 crc kubenswrapper[4799]: I1129 05:28:59.603395 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=10.603366886 podStartE2EDuration="10.603366886s" podCreationTimestamp="2025-11-29 05:28:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 05:28:59.588552422 +0000 UTC m=+3015.231482822" watchObservedRunningTime="2025-11-29 05:28:59.603366886 +0000 UTC m=+3015.246297286" Nov 29 05:28:59 crc kubenswrapper[4799]: I1129 05:28:59.614000 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-598444b44c-v5znr" podStartSLOduration=2.980416776 podStartE2EDuration="16.613977419s" podCreationTimestamp="2025-11-29 05:28:43 +0000 UTC" firstStartedPulling="2025-11-29 05:28:44.394712507 +0000 UTC m=+3000.037642907" lastFinishedPulling="2025-11-29 05:28:58.02827315 +0000 UTC m=+3013.671203550" observedRunningTime="2025-11-29 05:28:59.61110647 +0000 UTC m=+3015.254036870" watchObservedRunningTime="2025-11-29 05:28:59.613977419 +0000 UTC m=+3015.256907809" Nov 29 05:28:59 crc kubenswrapper[4799]: I1129 05:28:59.633915 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7c5488f496-jvw7p" podStartSLOduration=3.125733522 podStartE2EDuration="13.633892945s" podCreationTimestamp="2025-11-29 05:28:46 +0000 UTC" firstStartedPulling="2025-11-29 05:28:47.52065443 +0000 UTC m=+3003.163584820" lastFinishedPulling="2025-11-29 05:28:58.028813843 +0000 UTC m=+3013.671744243" observedRunningTime="2025-11-29 05:28:59.633752461 +0000 UTC m=+3015.276682861" watchObservedRunningTime="2025-11-29 05:28:59.633892945 +0000 UTC m=+3015.276823345" Nov 29 05:29:00 crc kubenswrapper[4799]: I1129 05:29:00.351981 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 29 05:29:00 crc kubenswrapper[4799]: I1129 05:29:00.352055 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 29 05:29:00 crc kubenswrapper[4799]: I1129 05:29:00.405994 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 29 05:29:00 crc kubenswrapper[4799]: I1129 05:29:00.412348 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 29 05:29:00 crc kubenswrapper[4799]: I1129 05:29:00.562337 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 29 05:29:00 crc kubenswrapper[4799]: I1129 05:29:00.565003 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 29 05:29:02 crc kubenswrapper[4799]: I1129 05:29:02.165607 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 29 05:29:02 crc kubenswrapper[4799]: I1129 05:29:02.166282 4799 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 29 05:29:02 crc kubenswrapper[4799]: I1129 05:29:02.468554 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 29 05:29:02 crc kubenswrapper[4799]: I1129 05:29:02.591266 4799 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 29 05:29:03 crc kubenswrapper[4799]: I1129 05:29:03.253469 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 29 05:29:03 crc kubenswrapper[4799]: I1129 05:29:03.437633 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-566ff59745-jp249" Nov 29 05:29:03 crc kubenswrapper[4799]: I1129 05:29:03.693062 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-598444b44c-v5znr" Nov 29 05:29:05 crc kubenswrapper[4799]: I1129 05:29:05.746415 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 29 05:29:06 crc kubenswrapper[4799]: I1129 05:29:06.797980 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-86d9b546cb-5xxhb" Nov 29 05:29:06 crc kubenswrapper[4799]: I1129 05:29:06.798543 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-86d9b546cb-5xxhb" Nov 29 05:29:06 crc kubenswrapper[4799]: I1129 05:29:06.927225 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7c5488f496-jvw7p" Nov 29 05:29:06 crc kubenswrapper[4799]: I1129 05:29:06.927306 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7c5488f496-jvw7p" Nov 29 05:29:07 crc kubenswrapper[4799]: I1129 05:29:07.689733 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-jlhf6" event={"ID":"4a730664-a7b6-4cbd-b657-d6f8ef3803b8","Type":"ContainerStarted","Data":"5e6f69a4304cd3351419612f0eabb5b6a202da0cc0f49f1df47483f3354e83e4"} Nov 29 05:29:07 crc kubenswrapper[4799]: I1129 05:29:07.725569 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-sync-jlhf6" podStartSLOduration=7.295924633 podStartE2EDuration="14.725543503s" podCreationTimestamp="2025-11-29 05:28:53 +0000 UTC" firstStartedPulling="2025-11-29 05:28:58.523547829 +0000 UTC m=+3014.166478229" lastFinishedPulling="2025-11-29 05:29:05.953166699 +0000 UTC m=+3021.596097099" observedRunningTime="2025-11-29 05:29:07.712644836 +0000 UTC m=+3023.355575236" watchObservedRunningTime="2025-11-29 05:29:07.725543503 +0000 UTC m=+3023.368473903" Nov 29 05:29:08 crc kubenswrapper[4799]: I1129 05:29:08.663903 4799 scope.go:117] "RemoveContainer" containerID="e6d4b3d1d5b5e1eeb13508d67de3d3c3aac52e76337b55e53793185c0d001d82" Nov 29 05:29:08 crc kubenswrapper[4799]: E1129 05:29:08.664389 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:29:16 crc kubenswrapper[4799]: I1129 05:29:16.801828 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-86d9b546cb-5xxhb" podUID="f09087c5-87d5-4cd1-9a91-cbe48713d9e9" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.236:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.236:8443: connect: connection refused" Nov 29 05:29:16 crc kubenswrapper[4799]: I1129 05:29:16.930107 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7c5488f496-jvw7p" podUID="93fb8bea-4246-40da-b480-c980b628e2b9" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.237:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.237:8443: connect: connection refused" Nov 29 05:29:17 crc kubenswrapper[4799]: I1129 05:29:17.852888 4799 generic.go:334] "Generic (PLEG): container finished" podID="4a730664-a7b6-4cbd-b657-d6f8ef3803b8" containerID="5e6f69a4304cd3351419612f0eabb5b6a202da0cc0f49f1df47483f3354e83e4" exitCode=0 Nov 29 05:29:17 crc kubenswrapper[4799]: I1129 05:29:17.853018 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-jlhf6" event={"ID":"4a730664-a7b6-4cbd-b657-d6f8ef3803b8","Type":"ContainerDied","Data":"5e6f69a4304cd3351419612f0eabb5b6a202da0cc0f49f1df47483f3354e83e4"} Nov 29 05:29:19 crc kubenswrapper[4799]: I1129 05:29:19.333190 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-jlhf6" Nov 29 05:29:19 crc kubenswrapper[4799]: I1129 05:29:19.491286 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a730664-a7b6-4cbd-b657-d6f8ef3803b8-combined-ca-bundle\") pod \"4a730664-a7b6-4cbd-b657-d6f8ef3803b8\" (UID: \"4a730664-a7b6-4cbd-b657-d6f8ef3803b8\") " Nov 29 05:29:19 crc kubenswrapper[4799]: I1129 05:29:19.492377 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a730664-a7b6-4cbd-b657-d6f8ef3803b8-config-data\") pod \"4a730664-a7b6-4cbd-b657-d6f8ef3803b8\" (UID: \"4a730664-a7b6-4cbd-b657-d6f8ef3803b8\") " Nov 29 05:29:19 crc kubenswrapper[4799]: I1129 05:29:19.492612 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/4a730664-a7b6-4cbd-b657-d6f8ef3803b8-job-config-data\") pod \"4a730664-a7b6-4cbd-b657-d6f8ef3803b8\" (UID: \"4a730664-a7b6-4cbd-b657-d6f8ef3803b8\") " Nov 29 05:29:19 crc kubenswrapper[4799]: I1129 05:29:19.493313 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bjkj2\" (UniqueName: \"kubernetes.io/projected/4a730664-a7b6-4cbd-b657-d6f8ef3803b8-kube-api-access-bjkj2\") pod \"4a730664-a7b6-4cbd-b657-d6f8ef3803b8\" (UID: \"4a730664-a7b6-4cbd-b657-d6f8ef3803b8\") " Nov 29 05:29:19 crc kubenswrapper[4799]: I1129 05:29:19.501604 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a730664-a7b6-4cbd-b657-d6f8ef3803b8-job-config-data" (OuterVolumeSpecName: "job-config-data") pod "4a730664-a7b6-4cbd-b657-d6f8ef3803b8" (UID: "4a730664-a7b6-4cbd-b657-d6f8ef3803b8"). InnerVolumeSpecName "job-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:29:19 crc kubenswrapper[4799]: I1129 05:29:19.510397 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a730664-a7b6-4cbd-b657-d6f8ef3803b8-config-data" (OuterVolumeSpecName: "config-data") pod "4a730664-a7b6-4cbd-b657-d6f8ef3803b8" (UID: "4a730664-a7b6-4cbd-b657-d6f8ef3803b8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:29:19 crc kubenswrapper[4799]: I1129 05:29:19.516996 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a730664-a7b6-4cbd-b657-d6f8ef3803b8-kube-api-access-bjkj2" (OuterVolumeSpecName: "kube-api-access-bjkj2") pod "4a730664-a7b6-4cbd-b657-d6f8ef3803b8" (UID: "4a730664-a7b6-4cbd-b657-d6f8ef3803b8"). InnerVolumeSpecName "kube-api-access-bjkj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:29:19 crc kubenswrapper[4799]: I1129 05:29:19.535196 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a730664-a7b6-4cbd-b657-d6f8ef3803b8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4a730664-a7b6-4cbd-b657-d6f8ef3803b8" (UID: "4a730664-a7b6-4cbd-b657-d6f8ef3803b8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:29:19 crc kubenswrapper[4799]: I1129 05:29:19.596048 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a730664-a7b6-4cbd-b657-d6f8ef3803b8-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:19 crc kubenswrapper[4799]: I1129 05:29:19.596087 4799 reconciler_common.go:293] "Volume detached for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/4a730664-a7b6-4cbd-b657-d6f8ef3803b8-job-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:19 crc kubenswrapper[4799]: I1129 05:29:19.596102 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bjkj2\" (UniqueName: \"kubernetes.io/projected/4a730664-a7b6-4cbd-b657-d6f8ef3803b8-kube-api-access-bjkj2\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:19 crc kubenswrapper[4799]: I1129 05:29:19.596114 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a730664-a7b6-4cbd-b657-d6f8ef3803b8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:19 crc kubenswrapper[4799]: I1129 05:29:19.902991 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-jlhf6" event={"ID":"4a730664-a7b6-4cbd-b657-d6f8ef3803b8","Type":"ContainerDied","Data":"6ccf320d5f0be9ea6360f97ded6dff3de3ee6e7fe45cacc4a93c7a13de90db9d"} Nov 29 05:29:19 crc kubenswrapper[4799]: I1129 05:29:19.903358 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6ccf320d5f0be9ea6360f97ded6dff3de3ee6e7fe45cacc4a93c7a13de90db9d" Nov 29 05:29:19 crc kubenswrapper[4799]: I1129 05:29:19.903130 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-jlhf6" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.272294 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Nov 29 05:29:20 crc kubenswrapper[4799]: E1129 05:29:20.281518 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a730664-a7b6-4cbd-b657-d6f8ef3803b8" containerName="manila-db-sync" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.281570 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a730664-a7b6-4cbd-b657-d6f8ef3803b8" containerName="manila-db-sync" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.281828 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a730664-a7b6-4cbd-b657-d6f8ef3803b8" containerName="manila-db-sync" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.283109 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.287475 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.287851 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.287994 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scripts" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.288018 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-nngct" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.292297 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.294175 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.296539 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.310701 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.325356 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/73c23628-dfd6-4496-9091-c1fbe3b092f6-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"73c23628-dfd6-4496-9091-c1fbe3b092f6\") " pod="openstack/manila-share-share1-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.325407 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5d3d2b3e-1604-4813-9710-85c349aed800-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"5d3d2b3e-1604-4813-9710-85c349aed800\") " pod="openstack/manila-scheduler-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.325436 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/73c23628-dfd6-4496-9091-c1fbe3b092f6-scripts\") pod \"manila-share-share1-0\" (UID: \"73c23628-dfd6-4496-9091-c1fbe3b092f6\") " pod="openstack/manila-share-share1-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.325490 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d3d2b3e-1604-4813-9710-85c349aed800-scripts\") pod \"manila-scheduler-0\" (UID: \"5d3d2b3e-1604-4813-9710-85c349aed800\") " pod="openstack/manila-scheduler-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.325510 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5d3d2b3e-1604-4813-9710-85c349aed800-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"5d3d2b3e-1604-4813-9710-85c349aed800\") " pod="openstack/manila-scheduler-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.325545 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d3d2b3e-1604-4813-9710-85c349aed800-config-data\") pod \"manila-scheduler-0\" (UID: \"5d3d2b3e-1604-4813-9710-85c349aed800\") " pod="openstack/manila-scheduler-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.325580 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73c23628-dfd6-4496-9091-c1fbe3b092f6-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"73c23628-dfd6-4496-9091-c1fbe3b092f6\") " pod="openstack/manila-share-share1-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.325772 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvnfg\" (UniqueName: \"kubernetes.io/projected/73c23628-dfd6-4496-9091-c1fbe3b092f6-kube-api-access-pvnfg\") pod \"manila-share-share1-0\" (UID: \"73c23628-dfd6-4496-9091-c1fbe3b092f6\") " pod="openstack/manila-share-share1-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.325874 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/73c23628-dfd6-4496-9091-c1fbe3b092f6-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"73c23628-dfd6-4496-9091-c1fbe3b092f6\") " pod="openstack/manila-share-share1-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.325908 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73c23628-dfd6-4496-9091-c1fbe3b092f6-config-data\") pod \"manila-share-share1-0\" (UID: \"73c23628-dfd6-4496-9091-c1fbe3b092f6\") " pod="openstack/manila-share-share1-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.325948 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/73c23628-dfd6-4496-9091-c1fbe3b092f6-ceph\") pod \"manila-share-share1-0\" (UID: \"73c23628-dfd6-4496-9091-c1fbe3b092f6\") " pod="openstack/manila-share-share1-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.325981 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vw6fd\" (UniqueName: \"kubernetes.io/projected/5d3d2b3e-1604-4813-9710-85c349aed800-kube-api-access-vw6fd\") pod \"manila-scheduler-0\" (UID: \"5d3d2b3e-1604-4813-9710-85c349aed800\") " pod="openstack/manila-scheduler-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.326136 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d3d2b3e-1604-4813-9710-85c349aed800-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"5d3d2b3e-1604-4813-9710-85c349aed800\") " pod="openstack/manila-scheduler-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.326285 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/73c23628-dfd6-4496-9091-c1fbe3b092f6-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"73c23628-dfd6-4496-9091-c1fbe3b092f6\") " pod="openstack/manila-share-share1-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.336359 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.413697 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-76b5fdb995-cbg8h"] Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.415656 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76b5fdb995-cbg8h" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.429526 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d3d2b3e-1604-4813-9710-85c349aed800-config-data\") pod \"manila-scheduler-0\" (UID: \"5d3d2b3e-1604-4813-9710-85c349aed800\") " pod="openstack/manila-scheduler-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.429586 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/87fc92de-e3e3-454f-886a-226aa591c580-ovsdbserver-nb\") pod \"dnsmasq-dns-76b5fdb995-cbg8h\" (UID: \"87fc92de-e3e3-454f-886a-226aa591c580\") " pod="openstack/dnsmasq-dns-76b5fdb995-cbg8h" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.429620 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73c23628-dfd6-4496-9091-c1fbe3b092f6-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"73c23628-dfd6-4496-9091-c1fbe3b092f6\") " pod="openstack/manila-share-share1-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.429643 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/87fc92de-e3e3-454f-886a-226aa591c580-dns-svc\") pod \"dnsmasq-dns-76b5fdb995-cbg8h\" (UID: \"87fc92de-e3e3-454f-886a-226aa591c580\") " pod="openstack/dnsmasq-dns-76b5fdb995-cbg8h" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.429675 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvnfg\" (UniqueName: \"kubernetes.io/projected/73c23628-dfd6-4496-9091-c1fbe3b092f6-kube-api-access-pvnfg\") pod \"manila-share-share1-0\" (UID: \"73c23628-dfd6-4496-9091-c1fbe3b092f6\") " pod="openstack/manila-share-share1-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.429704 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/73c23628-dfd6-4496-9091-c1fbe3b092f6-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"73c23628-dfd6-4496-9091-c1fbe3b092f6\") " pod="openstack/manila-share-share1-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.429727 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73c23628-dfd6-4496-9091-c1fbe3b092f6-config-data\") pod \"manila-share-share1-0\" (UID: \"73c23628-dfd6-4496-9091-c1fbe3b092f6\") " pod="openstack/manila-share-share1-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.429747 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/73c23628-dfd6-4496-9091-c1fbe3b092f6-ceph\") pod \"manila-share-share1-0\" (UID: \"73c23628-dfd6-4496-9091-c1fbe3b092f6\") " pod="openstack/manila-share-share1-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.429766 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vw6fd\" (UniqueName: \"kubernetes.io/projected/5d3d2b3e-1604-4813-9710-85c349aed800-kube-api-access-vw6fd\") pod \"manila-scheduler-0\" (UID: \"5d3d2b3e-1604-4813-9710-85c349aed800\") " pod="openstack/manila-scheduler-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.429804 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d3d2b3e-1604-4813-9710-85c349aed800-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"5d3d2b3e-1604-4813-9710-85c349aed800\") " pod="openstack/manila-scheduler-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.429830 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6nr5\" (UniqueName: \"kubernetes.io/projected/87fc92de-e3e3-454f-886a-226aa591c580-kube-api-access-f6nr5\") pod \"dnsmasq-dns-76b5fdb995-cbg8h\" (UID: \"87fc92de-e3e3-454f-886a-226aa591c580\") " pod="openstack/dnsmasq-dns-76b5fdb995-cbg8h" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.429857 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87fc92de-e3e3-454f-886a-226aa591c580-config\") pod \"dnsmasq-dns-76b5fdb995-cbg8h\" (UID: \"87fc92de-e3e3-454f-886a-226aa591c580\") " pod="openstack/dnsmasq-dns-76b5fdb995-cbg8h" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.429880 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/73c23628-dfd6-4496-9091-c1fbe3b092f6-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"73c23628-dfd6-4496-9091-c1fbe3b092f6\") " pod="openstack/manila-share-share1-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.429907 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/73c23628-dfd6-4496-9091-c1fbe3b092f6-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"73c23628-dfd6-4496-9091-c1fbe3b092f6\") " pod="openstack/manila-share-share1-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.429947 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5d3d2b3e-1604-4813-9710-85c349aed800-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"5d3d2b3e-1604-4813-9710-85c349aed800\") " pod="openstack/manila-scheduler-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.429967 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/87fc92de-e3e3-454f-886a-226aa591c580-ovsdbserver-sb\") pod \"dnsmasq-dns-76b5fdb995-cbg8h\" (UID: \"87fc92de-e3e3-454f-886a-226aa591c580\") " pod="openstack/dnsmasq-dns-76b5fdb995-cbg8h" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.429988 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/73c23628-dfd6-4496-9091-c1fbe3b092f6-scripts\") pod \"manila-share-share1-0\" (UID: \"73c23628-dfd6-4496-9091-c1fbe3b092f6\") " pod="openstack/manila-share-share1-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.430003 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/87fc92de-e3e3-454f-886a-226aa591c580-openstack-edpm-ipam\") pod \"dnsmasq-dns-76b5fdb995-cbg8h\" (UID: \"87fc92de-e3e3-454f-886a-226aa591c580\") " pod="openstack/dnsmasq-dns-76b5fdb995-cbg8h" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.430059 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d3d2b3e-1604-4813-9710-85c349aed800-scripts\") pod \"manila-scheduler-0\" (UID: \"5d3d2b3e-1604-4813-9710-85c349aed800\") " pod="openstack/manila-scheduler-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.430079 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5d3d2b3e-1604-4813-9710-85c349aed800-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"5d3d2b3e-1604-4813-9710-85c349aed800\") " pod="openstack/manila-scheduler-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.431951 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/73c23628-dfd6-4496-9091-c1fbe3b092f6-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"73c23628-dfd6-4496-9091-c1fbe3b092f6\") " pod="openstack/manila-share-share1-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.432002 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/73c23628-dfd6-4496-9091-c1fbe3b092f6-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"73c23628-dfd6-4496-9091-c1fbe3b092f6\") " pod="openstack/manila-share-share1-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.436026 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d3d2b3e-1604-4813-9710-85c349aed800-config-data\") pod \"manila-scheduler-0\" (UID: \"5d3d2b3e-1604-4813-9710-85c349aed800\") " pod="openstack/manila-scheduler-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.436068 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5d3d2b3e-1604-4813-9710-85c349aed800-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"5d3d2b3e-1604-4813-9710-85c349aed800\") " pod="openstack/manila-scheduler-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.436917 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d3d2b3e-1604-4813-9710-85c349aed800-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"5d3d2b3e-1604-4813-9710-85c349aed800\") " pod="openstack/manila-scheduler-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.441822 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73c23628-dfd6-4496-9091-c1fbe3b092f6-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"73c23628-dfd6-4496-9091-c1fbe3b092f6\") " pod="openstack/manila-share-share1-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.441829 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/73c23628-dfd6-4496-9091-c1fbe3b092f6-scripts\") pod \"manila-share-share1-0\" (UID: \"73c23628-dfd6-4496-9091-c1fbe3b092f6\") " pod="openstack/manila-share-share1-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.441930 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5d3d2b3e-1604-4813-9710-85c349aed800-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"5d3d2b3e-1604-4813-9710-85c349aed800\") " pod="openstack/manila-scheduler-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.442849 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/73c23628-dfd6-4496-9091-c1fbe3b092f6-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"73c23628-dfd6-4496-9091-c1fbe3b092f6\") " pod="openstack/manila-share-share1-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.448262 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d3d2b3e-1604-4813-9710-85c349aed800-scripts\") pod \"manila-scheduler-0\" (UID: \"5d3d2b3e-1604-4813-9710-85c349aed800\") " pod="openstack/manila-scheduler-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.454408 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/73c23628-dfd6-4496-9091-c1fbe3b092f6-ceph\") pod \"manila-share-share1-0\" (UID: \"73c23628-dfd6-4496-9091-c1fbe3b092f6\") " pod="openstack/manila-share-share1-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.456893 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76b5fdb995-cbg8h"] Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.458887 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73c23628-dfd6-4496-9091-c1fbe3b092f6-config-data\") pod \"manila-share-share1-0\" (UID: \"73c23628-dfd6-4496-9091-c1fbe3b092f6\") " pod="openstack/manila-share-share1-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.474430 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvnfg\" (UniqueName: \"kubernetes.io/projected/73c23628-dfd6-4496-9091-c1fbe3b092f6-kube-api-access-pvnfg\") pod \"manila-share-share1-0\" (UID: \"73c23628-dfd6-4496-9091-c1fbe3b092f6\") " pod="openstack/manila-share-share1-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.476760 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vw6fd\" (UniqueName: \"kubernetes.io/projected/5d3d2b3e-1604-4813-9710-85c349aed800-kube-api-access-vw6fd\") pod \"manila-scheduler-0\" (UID: \"5d3d2b3e-1604-4813-9710-85c349aed800\") " pod="openstack/manila-scheduler-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.533554 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/87fc92de-e3e3-454f-886a-226aa591c580-dns-svc\") pod \"dnsmasq-dns-76b5fdb995-cbg8h\" (UID: \"87fc92de-e3e3-454f-886a-226aa591c580\") " pod="openstack/dnsmasq-dns-76b5fdb995-cbg8h" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.533659 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6nr5\" (UniqueName: \"kubernetes.io/projected/87fc92de-e3e3-454f-886a-226aa591c580-kube-api-access-f6nr5\") pod \"dnsmasq-dns-76b5fdb995-cbg8h\" (UID: \"87fc92de-e3e3-454f-886a-226aa591c580\") " pod="openstack/dnsmasq-dns-76b5fdb995-cbg8h" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.533692 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87fc92de-e3e3-454f-886a-226aa591c580-config\") pod \"dnsmasq-dns-76b5fdb995-cbg8h\" (UID: \"87fc92de-e3e3-454f-886a-226aa591c580\") " pod="openstack/dnsmasq-dns-76b5fdb995-cbg8h" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.533731 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/87fc92de-e3e3-454f-886a-226aa591c580-ovsdbserver-sb\") pod \"dnsmasq-dns-76b5fdb995-cbg8h\" (UID: \"87fc92de-e3e3-454f-886a-226aa591c580\") " pod="openstack/dnsmasq-dns-76b5fdb995-cbg8h" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.533754 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/87fc92de-e3e3-454f-886a-226aa591c580-openstack-edpm-ipam\") pod \"dnsmasq-dns-76b5fdb995-cbg8h\" (UID: \"87fc92de-e3e3-454f-886a-226aa591c580\") " pod="openstack/dnsmasq-dns-76b5fdb995-cbg8h" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.533843 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/87fc92de-e3e3-454f-886a-226aa591c580-ovsdbserver-nb\") pod \"dnsmasq-dns-76b5fdb995-cbg8h\" (UID: \"87fc92de-e3e3-454f-886a-226aa591c580\") " pod="openstack/dnsmasq-dns-76b5fdb995-cbg8h" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.536554 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87fc92de-e3e3-454f-886a-226aa591c580-config\") pod \"dnsmasq-dns-76b5fdb995-cbg8h\" (UID: \"87fc92de-e3e3-454f-886a-226aa591c580\") " pod="openstack/dnsmasq-dns-76b5fdb995-cbg8h" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.537192 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/87fc92de-e3e3-454f-886a-226aa591c580-dns-svc\") pod \"dnsmasq-dns-76b5fdb995-cbg8h\" (UID: \"87fc92de-e3e3-454f-886a-226aa591c580\") " pod="openstack/dnsmasq-dns-76b5fdb995-cbg8h" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.538047 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/87fc92de-e3e3-454f-886a-226aa591c580-ovsdbserver-sb\") pod \"dnsmasq-dns-76b5fdb995-cbg8h\" (UID: \"87fc92de-e3e3-454f-886a-226aa591c580\") " pod="openstack/dnsmasq-dns-76b5fdb995-cbg8h" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.538358 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/87fc92de-e3e3-454f-886a-226aa591c580-ovsdbserver-nb\") pod \"dnsmasq-dns-76b5fdb995-cbg8h\" (UID: \"87fc92de-e3e3-454f-886a-226aa591c580\") " pod="openstack/dnsmasq-dns-76b5fdb995-cbg8h" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.538698 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/87fc92de-e3e3-454f-886a-226aa591c580-openstack-edpm-ipam\") pod \"dnsmasq-dns-76b5fdb995-cbg8h\" (UID: \"87fc92de-e3e3-454f-886a-226aa591c580\") " pod="openstack/dnsmasq-dns-76b5fdb995-cbg8h" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.567844 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6nr5\" (UniqueName: \"kubernetes.io/projected/87fc92de-e3e3-454f-886a-226aa591c580-kube-api-access-f6nr5\") pod \"dnsmasq-dns-76b5fdb995-cbg8h\" (UID: \"87fc92de-e3e3-454f-886a-226aa591c580\") " pod="openstack/dnsmasq-dns-76b5fdb995-cbg8h" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.611840 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.616990 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.622273 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.627374 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.635281 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62bb2338-9e2b-4f62-8652-238fdd8063b7-scripts\") pod \"manila-api-0\" (UID: \"62bb2338-9e2b-4f62-8652-238fdd8063b7\") " pod="openstack/manila-api-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.635332 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65n5h\" (UniqueName: \"kubernetes.io/projected/62bb2338-9e2b-4f62-8652-238fdd8063b7-kube-api-access-65n5h\") pod \"manila-api-0\" (UID: \"62bb2338-9e2b-4f62-8652-238fdd8063b7\") " pod="openstack/manila-api-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.635435 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62bb2338-9e2b-4f62-8652-238fdd8063b7-logs\") pod \"manila-api-0\" (UID: \"62bb2338-9e2b-4f62-8652-238fdd8063b7\") " pod="openstack/manila-api-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.635463 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/62bb2338-9e2b-4f62-8652-238fdd8063b7-config-data-custom\") pod \"manila-api-0\" (UID: \"62bb2338-9e2b-4f62-8652-238fdd8063b7\") " pod="openstack/manila-api-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.635495 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/62bb2338-9e2b-4f62-8652-238fdd8063b7-etc-machine-id\") pod \"manila-api-0\" (UID: \"62bb2338-9e2b-4f62-8652-238fdd8063b7\") " pod="openstack/manila-api-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.635566 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62bb2338-9e2b-4f62-8652-238fdd8063b7-config-data\") pod \"manila-api-0\" (UID: \"62bb2338-9e2b-4f62-8652-238fdd8063b7\") " pod="openstack/manila-api-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.635617 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62bb2338-9e2b-4f62-8652-238fdd8063b7-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"62bb2338-9e2b-4f62-8652-238fdd8063b7\") " pod="openstack/manila-api-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.636157 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.651119 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.740101 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/62bb2338-9e2b-4f62-8652-238fdd8063b7-etc-machine-id\") pod \"manila-api-0\" (UID: \"62bb2338-9e2b-4f62-8652-238fdd8063b7\") " pod="openstack/manila-api-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.740439 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/62bb2338-9e2b-4f62-8652-238fdd8063b7-etc-machine-id\") pod \"manila-api-0\" (UID: \"62bb2338-9e2b-4f62-8652-238fdd8063b7\") " pod="openstack/manila-api-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.745767 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62bb2338-9e2b-4f62-8652-238fdd8063b7-config-data\") pod \"manila-api-0\" (UID: \"62bb2338-9e2b-4f62-8652-238fdd8063b7\") " pod="openstack/manila-api-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.746251 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62bb2338-9e2b-4f62-8652-238fdd8063b7-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"62bb2338-9e2b-4f62-8652-238fdd8063b7\") " pod="openstack/manila-api-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.750266 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62bb2338-9e2b-4f62-8652-238fdd8063b7-scripts\") pod \"manila-api-0\" (UID: \"62bb2338-9e2b-4f62-8652-238fdd8063b7\") " pod="openstack/manila-api-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.750324 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65n5h\" (UniqueName: \"kubernetes.io/projected/62bb2338-9e2b-4f62-8652-238fdd8063b7-kube-api-access-65n5h\") pod \"manila-api-0\" (UID: \"62bb2338-9e2b-4f62-8652-238fdd8063b7\") " pod="openstack/manila-api-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.750552 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62bb2338-9e2b-4f62-8652-238fdd8063b7-logs\") pod \"manila-api-0\" (UID: \"62bb2338-9e2b-4f62-8652-238fdd8063b7\") " pod="openstack/manila-api-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.750607 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/62bb2338-9e2b-4f62-8652-238fdd8063b7-config-data-custom\") pod \"manila-api-0\" (UID: \"62bb2338-9e2b-4f62-8652-238fdd8063b7\") " pod="openstack/manila-api-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.755264 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62bb2338-9e2b-4f62-8652-238fdd8063b7-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"62bb2338-9e2b-4f62-8652-238fdd8063b7\") " pod="openstack/manila-api-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.761769 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62bb2338-9e2b-4f62-8652-238fdd8063b7-logs\") pod \"manila-api-0\" (UID: \"62bb2338-9e2b-4f62-8652-238fdd8063b7\") " pod="openstack/manila-api-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.796841 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/62bb2338-9e2b-4f62-8652-238fdd8063b7-config-data-custom\") pod \"manila-api-0\" (UID: \"62bb2338-9e2b-4f62-8652-238fdd8063b7\") " pod="openstack/manila-api-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.797334 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62bb2338-9e2b-4f62-8652-238fdd8063b7-scripts\") pod \"manila-api-0\" (UID: \"62bb2338-9e2b-4f62-8652-238fdd8063b7\") " pod="openstack/manila-api-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.797532 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65n5h\" (UniqueName: \"kubernetes.io/projected/62bb2338-9e2b-4f62-8652-238fdd8063b7-kube-api-access-65n5h\") pod \"manila-api-0\" (UID: \"62bb2338-9e2b-4f62-8652-238fdd8063b7\") " pod="openstack/manila-api-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.797614 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62bb2338-9e2b-4f62-8652-238fdd8063b7-config-data\") pod \"manila-api-0\" (UID: \"62bb2338-9e2b-4f62-8652-238fdd8063b7\") " pod="openstack/manila-api-0" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.869069 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76b5fdb995-cbg8h" Nov 29 05:29:20 crc kubenswrapper[4799]: I1129 05:29:20.961674 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Nov 29 05:29:21 crc kubenswrapper[4799]: I1129 05:29:21.323777 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Nov 29 05:29:21 crc kubenswrapper[4799]: I1129 05:29:21.557623 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Nov 29 05:29:21 crc kubenswrapper[4799]: W1129 05:29:21.565017 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod73c23628_dfd6_4496_9091_c1fbe3b092f6.slice/crio-13d48ed228ea43f445daa4c3f940ab7e84bfeef23a264fe4e38abaa582bcf9c6 WatchSource:0}: Error finding container 13d48ed228ea43f445daa4c3f940ab7e84bfeef23a264fe4e38abaa582bcf9c6: Status 404 returned error can't find the container with id 13d48ed228ea43f445daa4c3f940ab7e84bfeef23a264fe4e38abaa582bcf9c6 Nov 29 05:29:21 crc kubenswrapper[4799]: I1129 05:29:21.604211 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76b5fdb995-cbg8h"] Nov 29 05:29:21 crc kubenswrapper[4799]: W1129 05:29:21.613206 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod87fc92de_e3e3_454f_886a_226aa591c580.slice/crio-174c40bbbdca55cc5aeb426f8f0edae093471c958a0d9a40ae523bd306907b73 WatchSource:0}: Error finding container 174c40bbbdca55cc5aeb426f8f0edae093471c958a0d9a40ae523bd306907b73: Status 404 returned error can't find the container with id 174c40bbbdca55cc5aeb426f8f0edae093471c958a0d9a40ae523bd306907b73 Nov 29 05:29:21 crc kubenswrapper[4799]: I1129 05:29:21.696474 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Nov 29 05:29:21 crc kubenswrapper[4799]: W1129 05:29:21.722281 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod62bb2338_9e2b_4f62_8652_238fdd8063b7.slice/crio-8cdbb74a1e72f9a480033766ec17f63be55e2fe410abf1161ce057fcbb923175 WatchSource:0}: Error finding container 8cdbb74a1e72f9a480033766ec17f63be55e2fe410abf1161ce057fcbb923175: Status 404 returned error can't find the container with id 8cdbb74a1e72f9a480033766ec17f63be55e2fe410abf1161ce057fcbb923175 Nov 29 05:29:21 crc kubenswrapper[4799]: I1129 05:29:21.962898 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"73c23628-dfd6-4496-9091-c1fbe3b092f6","Type":"ContainerStarted","Data":"13d48ed228ea43f445daa4c3f940ab7e84bfeef23a264fe4e38abaa582bcf9c6"} Nov 29 05:29:21 crc kubenswrapper[4799]: I1129 05:29:21.968952 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76b5fdb995-cbg8h" event={"ID":"87fc92de-e3e3-454f-886a-226aa591c580","Type":"ContainerStarted","Data":"174c40bbbdca55cc5aeb426f8f0edae093471c958a0d9a40ae523bd306907b73"} Nov 29 05:29:21 crc kubenswrapper[4799]: I1129 05:29:21.977550 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"5d3d2b3e-1604-4813-9710-85c349aed800","Type":"ContainerStarted","Data":"ca955360d61a44a4c5676feaad11990950059dab400f30479e9e54b38507551f"} Nov 29 05:29:21 crc kubenswrapper[4799]: I1129 05:29:21.987323 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"62bb2338-9e2b-4f62-8652-238fdd8063b7","Type":"ContainerStarted","Data":"8cdbb74a1e72f9a480033766ec17f63be55e2fe410abf1161ce057fcbb923175"} Nov 29 05:29:22 crc kubenswrapper[4799]: E1129 05:29:22.202986 4799 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod87fc92de_e3e3_454f_886a_226aa591c580.slice/crio-conmon-8bae8491e3becae26ec6d2da367d7b49cf9ed9919a2085cba8de554a8a525455.scope\": RecentStats: unable to find data in memory cache]" Nov 29 05:29:23 crc kubenswrapper[4799]: I1129 05:29:23.003053 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"62bb2338-9e2b-4f62-8652-238fdd8063b7","Type":"ContainerStarted","Data":"ac8cee2b0d8b73b8054fe43eca42d2ee452f3b2acbb44bd14072b0f031ba8b97"} Nov 29 05:29:23 crc kubenswrapper[4799]: I1129 05:29:23.006402 4799 generic.go:334] "Generic (PLEG): container finished" podID="87fc92de-e3e3-454f-886a-226aa591c580" containerID="8bae8491e3becae26ec6d2da367d7b49cf9ed9919a2085cba8de554a8a525455" exitCode=0 Nov 29 05:29:23 crc kubenswrapper[4799]: I1129 05:29:23.006493 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76b5fdb995-cbg8h" event={"ID":"87fc92de-e3e3-454f-886a-226aa591c580","Type":"ContainerDied","Data":"8bae8491e3becae26ec6d2da367d7b49cf9ed9919a2085cba8de554a8a525455"} Nov 29 05:29:23 crc kubenswrapper[4799]: I1129 05:29:23.008393 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"5d3d2b3e-1604-4813-9710-85c349aed800","Type":"ContainerStarted","Data":"263b1dd71a80edeb22f2516a46469a2f88ea33b9ae67c572d59f9e348056324a"} Nov 29 05:29:23 crc kubenswrapper[4799]: I1129 05:29:23.473847 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Nov 29 05:29:23 crc kubenswrapper[4799]: I1129 05:29:23.661001 4799 scope.go:117] "RemoveContainer" containerID="e6d4b3d1d5b5e1eeb13508d67de3d3c3aac52e76337b55e53793185c0d001d82" Nov 29 05:29:23 crc kubenswrapper[4799]: E1129 05:29:23.661234 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:29:24 crc kubenswrapper[4799]: I1129 05:29:24.021232 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"62bb2338-9e2b-4f62-8652-238fdd8063b7","Type":"ContainerStarted","Data":"c9d7ae91be5b663bac35a8f4d01f3ca4da49ccde7511c11f0ab43286d377f6da"} Nov 29 05:29:24 crc kubenswrapper[4799]: I1129 05:29:24.022766 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Nov 29 05:29:24 crc kubenswrapper[4799]: I1129 05:29:24.027392 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76b5fdb995-cbg8h" event={"ID":"87fc92de-e3e3-454f-886a-226aa591c580","Type":"ContainerStarted","Data":"9b62cbf4bc309ec03fa5b16ce1d9a8ab184fd521385bb067a4fae49a0d8d41df"} Nov 29 05:29:24 crc kubenswrapper[4799]: I1129 05:29:24.027565 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-76b5fdb995-cbg8h" Nov 29 05:29:24 crc kubenswrapper[4799]: I1129 05:29:24.029191 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"5d3d2b3e-1604-4813-9710-85c349aed800","Type":"ContainerStarted","Data":"6c7a37a0fbbee07d218aa186b9370ef13bb948300d408f55ec4b3a1e234fa574"} Nov 29 05:29:24 crc kubenswrapper[4799]: I1129 05:29:24.048518 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=4.048493574 podStartE2EDuration="4.048493574s" podCreationTimestamp="2025-11-29 05:29:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 05:29:24.041100068 +0000 UTC m=+3039.684030468" watchObservedRunningTime="2025-11-29 05:29:24.048493574 +0000 UTC m=+3039.691423974" Nov 29 05:29:24 crc kubenswrapper[4799]: I1129 05:29:24.077751 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=3.312313526 podStartE2EDuration="4.077729691s" podCreationTimestamp="2025-11-29 05:29:20 +0000 UTC" firstStartedPulling="2025-11-29 05:29:21.32944842 +0000 UTC m=+3036.972378820" lastFinishedPulling="2025-11-29 05:29:22.094864585 +0000 UTC m=+3037.737794985" observedRunningTime="2025-11-29 05:29:24.072052386 +0000 UTC m=+3039.714982796" watchObservedRunningTime="2025-11-29 05:29:24.077729691 +0000 UTC m=+3039.720660081" Nov 29 05:29:24 crc kubenswrapper[4799]: I1129 05:29:24.106393 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-76b5fdb995-cbg8h" podStartSLOduration=4.106369355 podStartE2EDuration="4.106369355s" podCreationTimestamp="2025-11-29 05:29:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 05:29:24.09653028 +0000 UTC m=+3039.739460680" watchObservedRunningTime="2025-11-29 05:29:24.106369355 +0000 UTC m=+3039.749299755" Nov 29 05:29:24 crc kubenswrapper[4799]: I1129 05:29:24.815645 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 29 05:29:24 crc kubenswrapper[4799]: I1129 05:29:24.816542 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9e378b38-1cb5-4f77-b9a4-e6cfce8da188" containerName="ceilometer-central-agent" containerID="cri-o://86e09030d07ac79226803571fd562d1b65201ff6597388493a2af883ba02cab8" gracePeriod=30 Nov 29 05:29:24 crc kubenswrapper[4799]: I1129 05:29:24.816630 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9e378b38-1cb5-4f77-b9a4-e6cfce8da188" containerName="sg-core" containerID="cri-o://adc14bfee676568c6a1b21f35b94c262a6fe0699e51bf4453d9d71b62855b2d8" gracePeriod=30 Nov 29 05:29:24 crc kubenswrapper[4799]: I1129 05:29:24.816680 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9e378b38-1cb5-4f77-b9a4-e6cfce8da188" containerName="ceilometer-notification-agent" containerID="cri-o://d09b5086c0cd8c7c5bb69705d115dec19ed06948186960f1aa1eafe1b7cd88a8" gracePeriod=30 Nov 29 05:29:24 crc kubenswrapper[4799]: I1129 05:29:24.816720 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9e378b38-1cb5-4f77-b9a4-e6cfce8da188" containerName="proxy-httpd" containerID="cri-o://be4294eae32e4ed3aef4ebf11a13fbb8fb8da2c722a9d820a58674f8e9a082dc" gracePeriod=30 Nov 29 05:29:25 crc kubenswrapper[4799]: I1129 05:29:25.047566 4799 generic.go:334] "Generic (PLEG): container finished" podID="9e378b38-1cb5-4f77-b9a4-e6cfce8da188" containerID="adc14bfee676568c6a1b21f35b94c262a6fe0699e51bf4453d9d71b62855b2d8" exitCode=2 Nov 29 05:29:25 crc kubenswrapper[4799]: I1129 05:29:25.047664 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9e378b38-1cb5-4f77-b9a4-e6cfce8da188","Type":"ContainerDied","Data":"adc14bfee676568c6a1b21f35b94c262a6fe0699e51bf4453d9d71b62855b2d8"} Nov 29 05:29:25 crc kubenswrapper[4799]: I1129 05:29:25.047834 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-api-0" podUID="62bb2338-9e2b-4f62-8652-238fdd8063b7" containerName="manila-api-log" containerID="cri-o://ac8cee2b0d8b73b8054fe43eca42d2ee452f3b2acbb44bd14072b0f031ba8b97" gracePeriod=30 Nov 29 05:29:25 crc kubenswrapper[4799]: I1129 05:29:25.047940 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-api-0" podUID="62bb2338-9e2b-4f62-8652-238fdd8063b7" containerName="manila-api" containerID="cri-o://c9d7ae91be5b663bac35a8f4d01f3ca4da49ccde7511c11f0ab43286d377f6da" gracePeriod=30 Nov 29 05:29:26 crc kubenswrapper[4799]: I1129 05:29:26.059330 4799 generic.go:334] "Generic (PLEG): container finished" podID="62bb2338-9e2b-4f62-8652-238fdd8063b7" containerID="c9d7ae91be5b663bac35a8f4d01f3ca4da49ccde7511c11f0ab43286d377f6da" exitCode=0 Nov 29 05:29:26 crc kubenswrapper[4799]: I1129 05:29:26.059777 4799 generic.go:334] "Generic (PLEG): container finished" podID="62bb2338-9e2b-4f62-8652-238fdd8063b7" containerID="ac8cee2b0d8b73b8054fe43eca42d2ee452f3b2acbb44bd14072b0f031ba8b97" exitCode=143 Nov 29 05:29:26 crc kubenswrapper[4799]: I1129 05:29:26.059396 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"62bb2338-9e2b-4f62-8652-238fdd8063b7","Type":"ContainerDied","Data":"c9d7ae91be5b663bac35a8f4d01f3ca4da49ccde7511c11f0ab43286d377f6da"} Nov 29 05:29:26 crc kubenswrapper[4799]: I1129 05:29:26.059920 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"62bb2338-9e2b-4f62-8652-238fdd8063b7","Type":"ContainerDied","Data":"ac8cee2b0d8b73b8054fe43eca42d2ee452f3b2acbb44bd14072b0f031ba8b97"} Nov 29 05:29:26 crc kubenswrapper[4799]: I1129 05:29:26.059937 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"62bb2338-9e2b-4f62-8652-238fdd8063b7","Type":"ContainerDied","Data":"8cdbb74a1e72f9a480033766ec17f63be55e2fe410abf1161ce057fcbb923175"} Nov 29 05:29:26 crc kubenswrapper[4799]: I1129 05:29:26.059950 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8cdbb74a1e72f9a480033766ec17f63be55e2fe410abf1161ce057fcbb923175" Nov 29 05:29:26 crc kubenswrapper[4799]: I1129 05:29:26.063602 4799 generic.go:334] "Generic (PLEG): container finished" podID="9e378b38-1cb5-4f77-b9a4-e6cfce8da188" containerID="be4294eae32e4ed3aef4ebf11a13fbb8fb8da2c722a9d820a58674f8e9a082dc" exitCode=0 Nov 29 05:29:26 crc kubenswrapper[4799]: I1129 05:29:26.063637 4799 generic.go:334] "Generic (PLEG): container finished" podID="9e378b38-1cb5-4f77-b9a4-e6cfce8da188" containerID="86e09030d07ac79226803571fd562d1b65201ff6597388493a2af883ba02cab8" exitCode=0 Nov 29 05:29:26 crc kubenswrapper[4799]: I1129 05:29:26.063660 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9e378b38-1cb5-4f77-b9a4-e6cfce8da188","Type":"ContainerDied","Data":"be4294eae32e4ed3aef4ebf11a13fbb8fb8da2c722a9d820a58674f8e9a082dc"} Nov 29 05:29:26 crc kubenswrapper[4799]: I1129 05:29:26.063681 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9e378b38-1cb5-4f77-b9a4-e6cfce8da188","Type":"ContainerDied","Data":"86e09030d07ac79226803571fd562d1b65201ff6597388493a2af883ba02cab8"} Nov 29 05:29:26 crc kubenswrapper[4799]: I1129 05:29:26.099951 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Nov 29 05:29:26 crc kubenswrapper[4799]: I1129 05:29:26.200776 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62bb2338-9e2b-4f62-8652-238fdd8063b7-logs\") pod \"62bb2338-9e2b-4f62-8652-238fdd8063b7\" (UID: \"62bb2338-9e2b-4f62-8652-238fdd8063b7\") " Nov 29 05:29:26 crc kubenswrapper[4799]: I1129 05:29:26.200894 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62bb2338-9e2b-4f62-8652-238fdd8063b7-config-data\") pod \"62bb2338-9e2b-4f62-8652-238fdd8063b7\" (UID: \"62bb2338-9e2b-4f62-8652-238fdd8063b7\") " Nov 29 05:29:26 crc kubenswrapper[4799]: I1129 05:29:26.200974 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/62bb2338-9e2b-4f62-8652-238fdd8063b7-etc-machine-id\") pod \"62bb2338-9e2b-4f62-8652-238fdd8063b7\" (UID: \"62bb2338-9e2b-4f62-8652-238fdd8063b7\") " Nov 29 05:29:26 crc kubenswrapper[4799]: I1129 05:29:26.201128 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/62bb2338-9e2b-4f62-8652-238fdd8063b7-config-data-custom\") pod \"62bb2338-9e2b-4f62-8652-238fdd8063b7\" (UID: \"62bb2338-9e2b-4f62-8652-238fdd8063b7\") " Nov 29 05:29:26 crc kubenswrapper[4799]: I1129 05:29:26.201161 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62bb2338-9e2b-4f62-8652-238fdd8063b7-scripts\") pod \"62bb2338-9e2b-4f62-8652-238fdd8063b7\" (UID: \"62bb2338-9e2b-4f62-8652-238fdd8063b7\") " Nov 29 05:29:26 crc kubenswrapper[4799]: I1129 05:29:26.201325 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-65n5h\" (UniqueName: \"kubernetes.io/projected/62bb2338-9e2b-4f62-8652-238fdd8063b7-kube-api-access-65n5h\") pod \"62bb2338-9e2b-4f62-8652-238fdd8063b7\" (UID: \"62bb2338-9e2b-4f62-8652-238fdd8063b7\") " Nov 29 05:29:26 crc kubenswrapper[4799]: I1129 05:29:26.201382 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62bb2338-9e2b-4f62-8652-238fdd8063b7-combined-ca-bundle\") pod \"62bb2338-9e2b-4f62-8652-238fdd8063b7\" (UID: \"62bb2338-9e2b-4f62-8652-238fdd8063b7\") " Nov 29 05:29:26 crc kubenswrapper[4799]: I1129 05:29:26.203053 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/62bb2338-9e2b-4f62-8652-238fdd8063b7-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "62bb2338-9e2b-4f62-8652-238fdd8063b7" (UID: "62bb2338-9e2b-4f62-8652-238fdd8063b7"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 05:29:26 crc kubenswrapper[4799]: I1129 05:29:26.203546 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62bb2338-9e2b-4f62-8652-238fdd8063b7-logs" (OuterVolumeSpecName: "logs") pod "62bb2338-9e2b-4f62-8652-238fdd8063b7" (UID: "62bb2338-9e2b-4f62-8652-238fdd8063b7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:29:26 crc kubenswrapper[4799]: I1129 05:29:26.210647 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62bb2338-9e2b-4f62-8652-238fdd8063b7-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "62bb2338-9e2b-4f62-8652-238fdd8063b7" (UID: "62bb2338-9e2b-4f62-8652-238fdd8063b7"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:29:26 crc kubenswrapper[4799]: I1129 05:29:26.211041 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62bb2338-9e2b-4f62-8652-238fdd8063b7-kube-api-access-65n5h" (OuterVolumeSpecName: "kube-api-access-65n5h") pod "62bb2338-9e2b-4f62-8652-238fdd8063b7" (UID: "62bb2338-9e2b-4f62-8652-238fdd8063b7"). InnerVolumeSpecName "kube-api-access-65n5h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:29:26 crc kubenswrapper[4799]: I1129 05:29:26.220387 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62bb2338-9e2b-4f62-8652-238fdd8063b7-scripts" (OuterVolumeSpecName: "scripts") pod "62bb2338-9e2b-4f62-8652-238fdd8063b7" (UID: "62bb2338-9e2b-4f62-8652-238fdd8063b7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:29:26 crc kubenswrapper[4799]: I1129 05:29:26.246293 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62bb2338-9e2b-4f62-8652-238fdd8063b7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "62bb2338-9e2b-4f62-8652-238fdd8063b7" (UID: "62bb2338-9e2b-4f62-8652-238fdd8063b7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:29:26 crc kubenswrapper[4799]: I1129 05:29:26.274720 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62bb2338-9e2b-4f62-8652-238fdd8063b7-config-data" (OuterVolumeSpecName: "config-data") pod "62bb2338-9e2b-4f62-8652-238fdd8063b7" (UID: "62bb2338-9e2b-4f62-8652-238fdd8063b7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:29:26 crc kubenswrapper[4799]: I1129 05:29:26.304070 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62bb2338-9e2b-4f62-8652-238fdd8063b7-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:26 crc kubenswrapper[4799]: I1129 05:29:26.304143 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-65n5h\" (UniqueName: \"kubernetes.io/projected/62bb2338-9e2b-4f62-8652-238fdd8063b7-kube-api-access-65n5h\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:26 crc kubenswrapper[4799]: I1129 05:29:26.304160 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62bb2338-9e2b-4f62-8652-238fdd8063b7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:26 crc kubenswrapper[4799]: I1129 05:29:26.304170 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62bb2338-9e2b-4f62-8652-238fdd8063b7-logs\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:26 crc kubenswrapper[4799]: I1129 05:29:26.304181 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62bb2338-9e2b-4f62-8652-238fdd8063b7-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:26 crc kubenswrapper[4799]: I1129 05:29:26.304227 4799 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/62bb2338-9e2b-4f62-8652-238fdd8063b7-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:26 crc kubenswrapper[4799]: I1129 05:29:26.304236 4799 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/62bb2338-9e2b-4f62-8652-238fdd8063b7-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:27 crc kubenswrapper[4799]: I1129 05:29:27.127139 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Nov 29 05:29:27 crc kubenswrapper[4799]: I1129 05:29:27.180176 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Nov 29 05:29:27 crc kubenswrapper[4799]: I1129 05:29:27.222511 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-api-0"] Nov 29 05:29:27 crc kubenswrapper[4799]: I1129 05:29:27.247481 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Nov 29 05:29:27 crc kubenswrapper[4799]: E1129 05:29:27.248104 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62bb2338-9e2b-4f62-8652-238fdd8063b7" containerName="manila-api" Nov 29 05:29:27 crc kubenswrapper[4799]: I1129 05:29:27.248120 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="62bb2338-9e2b-4f62-8652-238fdd8063b7" containerName="manila-api" Nov 29 05:29:27 crc kubenswrapper[4799]: E1129 05:29:27.248142 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62bb2338-9e2b-4f62-8652-238fdd8063b7" containerName="manila-api-log" Nov 29 05:29:27 crc kubenswrapper[4799]: I1129 05:29:27.248149 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="62bb2338-9e2b-4f62-8652-238fdd8063b7" containerName="manila-api-log" Nov 29 05:29:27 crc kubenswrapper[4799]: I1129 05:29:27.248381 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="62bb2338-9e2b-4f62-8652-238fdd8063b7" containerName="manila-api-log" Nov 29 05:29:27 crc kubenswrapper[4799]: I1129 05:29:27.248408 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="62bb2338-9e2b-4f62-8652-238fdd8063b7" containerName="manila-api" Nov 29 05:29:27 crc kubenswrapper[4799]: I1129 05:29:27.249722 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Nov 29 05:29:27 crc kubenswrapper[4799]: I1129 05:29:27.252672 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Nov 29 05:29:27 crc kubenswrapper[4799]: I1129 05:29:27.253004 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-manila-internal-svc" Nov 29 05:29:27 crc kubenswrapper[4799]: I1129 05:29:27.253147 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-manila-public-svc" Nov 29 05:29:27 crc kubenswrapper[4799]: I1129 05:29:27.270222 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Nov 29 05:29:27 crc kubenswrapper[4799]: I1129 05:29:27.346633 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8b16394-158d-47e0-b406-18636e57be57-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"b8b16394-158d-47e0-b406-18636e57be57\") " pod="openstack/manila-api-0" Nov 29 05:29:27 crc kubenswrapper[4799]: I1129 05:29:27.347002 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8b16394-158d-47e0-b406-18636e57be57-scripts\") pod \"manila-api-0\" (UID: \"b8b16394-158d-47e0-b406-18636e57be57\") " pod="openstack/manila-api-0" Nov 29 05:29:27 crc kubenswrapper[4799]: I1129 05:29:27.347162 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b8b16394-158d-47e0-b406-18636e57be57-config-data-custom\") pod \"manila-api-0\" (UID: \"b8b16394-158d-47e0-b406-18636e57be57\") " pod="openstack/manila-api-0" Nov 29 05:29:27 crc kubenswrapper[4799]: I1129 05:29:27.347239 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8b16394-158d-47e0-b406-18636e57be57-config-data\") pod \"manila-api-0\" (UID: \"b8b16394-158d-47e0-b406-18636e57be57\") " pod="openstack/manila-api-0" Nov 29 05:29:27 crc kubenswrapper[4799]: I1129 05:29:27.347352 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b8b16394-158d-47e0-b406-18636e57be57-internal-tls-certs\") pod \"manila-api-0\" (UID: \"b8b16394-158d-47e0-b406-18636e57be57\") " pod="openstack/manila-api-0" Nov 29 05:29:27 crc kubenswrapper[4799]: I1129 05:29:27.347446 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvgms\" (UniqueName: \"kubernetes.io/projected/b8b16394-158d-47e0-b406-18636e57be57-kube-api-access-xvgms\") pod \"manila-api-0\" (UID: \"b8b16394-158d-47e0-b406-18636e57be57\") " pod="openstack/manila-api-0" Nov 29 05:29:27 crc kubenswrapper[4799]: I1129 05:29:27.347541 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b8b16394-158d-47e0-b406-18636e57be57-logs\") pod \"manila-api-0\" (UID: \"b8b16394-158d-47e0-b406-18636e57be57\") " pod="openstack/manila-api-0" Nov 29 05:29:27 crc kubenswrapper[4799]: I1129 05:29:27.347618 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b8b16394-158d-47e0-b406-18636e57be57-public-tls-certs\") pod \"manila-api-0\" (UID: \"b8b16394-158d-47e0-b406-18636e57be57\") " pod="openstack/manila-api-0" Nov 29 05:29:27 crc kubenswrapper[4799]: I1129 05:29:27.347817 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b8b16394-158d-47e0-b406-18636e57be57-etc-machine-id\") pod \"manila-api-0\" (UID: \"b8b16394-158d-47e0-b406-18636e57be57\") " pod="openstack/manila-api-0" Nov 29 05:29:27 crc kubenswrapper[4799]: I1129 05:29:27.450115 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b8b16394-158d-47e0-b406-18636e57be57-config-data-custom\") pod \"manila-api-0\" (UID: \"b8b16394-158d-47e0-b406-18636e57be57\") " pod="openstack/manila-api-0" Nov 29 05:29:27 crc kubenswrapper[4799]: I1129 05:29:27.450163 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8b16394-158d-47e0-b406-18636e57be57-config-data\") pod \"manila-api-0\" (UID: \"b8b16394-158d-47e0-b406-18636e57be57\") " pod="openstack/manila-api-0" Nov 29 05:29:27 crc kubenswrapper[4799]: I1129 05:29:27.450227 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b8b16394-158d-47e0-b406-18636e57be57-internal-tls-certs\") pod \"manila-api-0\" (UID: \"b8b16394-158d-47e0-b406-18636e57be57\") " pod="openstack/manila-api-0" Nov 29 05:29:27 crc kubenswrapper[4799]: I1129 05:29:27.450261 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvgms\" (UniqueName: \"kubernetes.io/projected/b8b16394-158d-47e0-b406-18636e57be57-kube-api-access-xvgms\") pod \"manila-api-0\" (UID: \"b8b16394-158d-47e0-b406-18636e57be57\") " pod="openstack/manila-api-0" Nov 29 05:29:27 crc kubenswrapper[4799]: I1129 05:29:27.450301 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b8b16394-158d-47e0-b406-18636e57be57-logs\") pod \"manila-api-0\" (UID: \"b8b16394-158d-47e0-b406-18636e57be57\") " pod="openstack/manila-api-0" Nov 29 05:29:27 crc kubenswrapper[4799]: I1129 05:29:27.450324 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b8b16394-158d-47e0-b406-18636e57be57-public-tls-certs\") pod \"manila-api-0\" (UID: \"b8b16394-158d-47e0-b406-18636e57be57\") " pod="openstack/manila-api-0" Nov 29 05:29:27 crc kubenswrapper[4799]: I1129 05:29:27.450359 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b8b16394-158d-47e0-b406-18636e57be57-etc-machine-id\") pod \"manila-api-0\" (UID: \"b8b16394-158d-47e0-b406-18636e57be57\") " pod="openstack/manila-api-0" Nov 29 05:29:27 crc kubenswrapper[4799]: I1129 05:29:27.450375 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8b16394-158d-47e0-b406-18636e57be57-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"b8b16394-158d-47e0-b406-18636e57be57\") " pod="openstack/manila-api-0" Nov 29 05:29:27 crc kubenswrapper[4799]: I1129 05:29:27.450395 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8b16394-158d-47e0-b406-18636e57be57-scripts\") pod \"manila-api-0\" (UID: \"b8b16394-158d-47e0-b406-18636e57be57\") " pod="openstack/manila-api-0" Nov 29 05:29:27 crc kubenswrapper[4799]: I1129 05:29:27.453003 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b8b16394-158d-47e0-b406-18636e57be57-logs\") pod \"manila-api-0\" (UID: \"b8b16394-158d-47e0-b406-18636e57be57\") " pod="openstack/manila-api-0" Nov 29 05:29:27 crc kubenswrapper[4799]: I1129 05:29:27.453437 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b8b16394-158d-47e0-b406-18636e57be57-etc-machine-id\") pod \"manila-api-0\" (UID: \"b8b16394-158d-47e0-b406-18636e57be57\") " pod="openstack/manila-api-0" Nov 29 05:29:27 crc kubenswrapper[4799]: I1129 05:29:27.456774 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8b16394-158d-47e0-b406-18636e57be57-scripts\") pod \"manila-api-0\" (UID: \"b8b16394-158d-47e0-b406-18636e57be57\") " pod="openstack/manila-api-0" Nov 29 05:29:27 crc kubenswrapper[4799]: I1129 05:29:27.459947 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b8b16394-158d-47e0-b406-18636e57be57-internal-tls-certs\") pod \"manila-api-0\" (UID: \"b8b16394-158d-47e0-b406-18636e57be57\") " pod="openstack/manila-api-0" Nov 29 05:29:27 crc kubenswrapper[4799]: I1129 05:29:27.462580 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8b16394-158d-47e0-b406-18636e57be57-config-data\") pod \"manila-api-0\" (UID: \"b8b16394-158d-47e0-b406-18636e57be57\") " pod="openstack/manila-api-0" Nov 29 05:29:27 crc kubenswrapper[4799]: I1129 05:29:27.463120 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8b16394-158d-47e0-b406-18636e57be57-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"b8b16394-158d-47e0-b406-18636e57be57\") " pod="openstack/manila-api-0" Nov 29 05:29:27 crc kubenswrapper[4799]: I1129 05:29:27.470593 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b8b16394-158d-47e0-b406-18636e57be57-public-tls-certs\") pod \"manila-api-0\" (UID: \"b8b16394-158d-47e0-b406-18636e57be57\") " pod="openstack/manila-api-0" Nov 29 05:29:27 crc kubenswrapper[4799]: I1129 05:29:27.471384 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvgms\" (UniqueName: \"kubernetes.io/projected/b8b16394-158d-47e0-b406-18636e57be57-kube-api-access-xvgms\") pod \"manila-api-0\" (UID: \"b8b16394-158d-47e0-b406-18636e57be57\") " pod="openstack/manila-api-0" Nov 29 05:29:27 crc kubenswrapper[4799]: I1129 05:29:27.475219 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b8b16394-158d-47e0-b406-18636e57be57-config-data-custom\") pod \"manila-api-0\" (UID: \"b8b16394-158d-47e0-b406-18636e57be57\") " pod="openstack/manila-api-0" Nov 29 05:29:27 crc kubenswrapper[4799]: I1129 05:29:27.588881 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Nov 29 05:29:28 crc kubenswrapper[4799]: I1129 05:29:28.670751 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62bb2338-9e2b-4f62-8652-238fdd8063b7" path="/var/lib/kubelet/pods/62bb2338-9e2b-4f62-8652-238fdd8063b7/volumes" Nov 29 05:29:28 crc kubenswrapper[4799]: I1129 05:29:28.711505 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="9e378b38-1cb5-4f77-b9a4-e6cfce8da188" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.180:3000/\": dial tcp 10.217.0.180:3000: connect: connection refused" Nov 29 05:29:29 crc kubenswrapper[4799]: I1129 05:29:29.072252 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-86d9b546cb-5xxhb" Nov 29 05:29:29 crc kubenswrapper[4799]: I1129 05:29:29.091616 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-7c5488f496-jvw7p" Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.169466 4799 generic.go:334] "Generic (PLEG): container finished" podID="9e378b38-1cb5-4f77-b9a4-e6cfce8da188" containerID="d09b5086c0cd8c7c5bb69705d115dec19ed06948186960f1aa1eafe1b7cd88a8" exitCode=0 Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.169611 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9e378b38-1cb5-4f77-b9a4-e6cfce8da188","Type":"ContainerDied","Data":"d09b5086c0cd8c7c5bb69705d115dec19ed06948186960f1aa1eafe1b7cd88a8"} Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.176088 4799 generic.go:334] "Generic (PLEG): container finished" podID="04de4a38-5d09-4c68-b7a2-be6482748e50" containerID="f8801fb2760fa321e80c168af0997ee85b677223a66f5991257e83509f823137" exitCode=137 Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.176114 4799 generic.go:334] "Generic (PLEG): container finished" podID="04de4a38-5d09-4c68-b7a2-be6482748e50" containerID="69cac5b4704d7a992b979fd8a84a6d26615f488bf691883048255f969711a221" exitCode=137 Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.176165 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-598444b44c-v5znr" event={"ID":"04de4a38-5d09-4c68-b7a2-be6482748e50","Type":"ContainerDied","Data":"f8801fb2760fa321e80c168af0997ee85b677223a66f5991257e83509f823137"} Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.176198 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-598444b44c-v5znr" event={"ID":"04de4a38-5d09-4c68-b7a2-be6482748e50","Type":"ContainerDied","Data":"69cac5b4704d7a992b979fd8a84a6d26615f488bf691883048255f969711a221"} Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.188849 4799 generic.go:334] "Generic (PLEG): container finished" podID="ebcbc04a-230e-46b1-be0d-af9ece769369" containerID="72186f35f25c7e3357a6a3815736e28ff75be3b34cc5516c50542e4d500121dc" exitCode=137 Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.188890 4799 generic.go:334] "Generic (PLEG): container finished" podID="ebcbc04a-230e-46b1-be0d-af9ece769369" containerID="a8ed286597384b45acd04d7e67e5c08d350f935fef4e4fbb30c795f46c0acf50" exitCode=137 Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.188923 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-566ff59745-jp249" event={"ID":"ebcbc04a-230e-46b1-be0d-af9ece769369","Type":"ContainerDied","Data":"72186f35f25c7e3357a6a3815736e28ff75be3b34cc5516c50542e4d500121dc"} Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.188961 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-566ff59745-jp249" event={"ID":"ebcbc04a-230e-46b1-be0d-af9ece769369","Type":"ContainerDied","Data":"a8ed286597384b45acd04d7e67e5c08d350f935fef4e4fbb30c795f46c0acf50"} Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.283558 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-598444b44c-v5znr" Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.418439 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-566ff59745-jp249" Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.430628 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/04de4a38-5d09-4c68-b7a2-be6482748e50-config-data\") pod \"04de4a38-5d09-4c68-b7a2-be6482748e50\" (UID: \"04de4a38-5d09-4c68-b7a2-be6482748e50\") " Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.430726 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/04de4a38-5d09-4c68-b7a2-be6482748e50-scripts\") pod \"04de4a38-5d09-4c68-b7a2-be6482748e50\" (UID: \"04de4a38-5d09-4c68-b7a2-be6482748e50\") " Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.430774 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rzbx\" (UniqueName: \"kubernetes.io/projected/04de4a38-5d09-4c68-b7a2-be6482748e50-kube-api-access-8rzbx\") pod \"04de4a38-5d09-4c68-b7a2-be6482748e50\" (UID: \"04de4a38-5d09-4c68-b7a2-be6482748e50\") " Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.430987 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04de4a38-5d09-4c68-b7a2-be6482748e50-logs\") pod \"04de4a38-5d09-4c68-b7a2-be6482748e50\" (UID: \"04de4a38-5d09-4c68-b7a2-be6482748e50\") " Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.431073 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/04de4a38-5d09-4c68-b7a2-be6482748e50-horizon-secret-key\") pod \"04de4a38-5d09-4c68-b7a2-be6482748e50\" (UID: \"04de4a38-5d09-4c68-b7a2-be6482748e50\") " Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.432074 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04de4a38-5d09-4c68-b7a2-be6482748e50-logs" (OuterVolumeSpecName: "logs") pod "04de4a38-5d09-4c68-b7a2-be6482748e50" (UID: "04de4a38-5d09-4c68-b7a2-be6482748e50"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.459252 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04de4a38-5d09-4c68-b7a2-be6482748e50-kube-api-access-8rzbx" (OuterVolumeSpecName: "kube-api-access-8rzbx") pod "04de4a38-5d09-4c68-b7a2-be6482748e50" (UID: "04de4a38-5d09-4c68-b7a2-be6482748e50"). InnerVolumeSpecName "kube-api-access-8rzbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.459326 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04de4a38-5d09-4c68-b7a2-be6482748e50-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "04de4a38-5d09-4c68-b7a2-be6482748e50" (UID: "04de4a38-5d09-4c68-b7a2-be6482748e50"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.481257 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04de4a38-5d09-4c68-b7a2-be6482748e50-scripts" (OuterVolumeSpecName: "scripts") pod "04de4a38-5d09-4c68-b7a2-be6482748e50" (UID: "04de4a38-5d09-4c68-b7a2-be6482748e50"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.486679 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04de4a38-5d09-4c68-b7a2-be6482748e50-config-data" (OuterVolumeSpecName: "config-data") pod "04de4a38-5d09-4c68-b7a2-be6482748e50" (UID: "04de4a38-5d09-4c68-b7a2-be6482748e50"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.533927 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ebcbc04a-230e-46b1-be0d-af9ece769369-scripts\") pod \"ebcbc04a-230e-46b1-be0d-af9ece769369\" (UID: \"ebcbc04a-230e-46b1-be0d-af9ece769369\") " Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.534120 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ebcbc04a-230e-46b1-be0d-af9ece769369-horizon-secret-key\") pod \"ebcbc04a-230e-46b1-be0d-af9ece769369\" (UID: \"ebcbc04a-230e-46b1-be0d-af9ece769369\") " Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.534227 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ebcbc04a-230e-46b1-be0d-af9ece769369-config-data\") pod \"ebcbc04a-230e-46b1-be0d-af9ece769369\" (UID: \"ebcbc04a-230e-46b1-be0d-af9ece769369\") " Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.534404 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ls4wl\" (UniqueName: \"kubernetes.io/projected/ebcbc04a-230e-46b1-be0d-af9ece769369-kube-api-access-ls4wl\") pod \"ebcbc04a-230e-46b1-be0d-af9ece769369\" (UID: \"ebcbc04a-230e-46b1-be0d-af9ece769369\") " Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.535021 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ebcbc04a-230e-46b1-be0d-af9ece769369-logs\") pod \"ebcbc04a-230e-46b1-be0d-af9ece769369\" (UID: \"ebcbc04a-230e-46b1-be0d-af9ece769369\") " Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.535732 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ebcbc04a-230e-46b1-be0d-af9ece769369-logs" (OuterVolumeSpecName: "logs") pod "ebcbc04a-230e-46b1-be0d-af9ece769369" (UID: "ebcbc04a-230e-46b1-be0d-af9ece769369"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.537170 4799 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/04de4a38-5d09-4c68-b7a2-be6482748e50-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.537263 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ebcbc04a-230e-46b1-be0d-af9ece769369-logs\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.537281 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/04de4a38-5d09-4c68-b7a2-be6482748e50-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.537319 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/04de4a38-5d09-4c68-b7a2-be6482748e50-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.537334 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rzbx\" (UniqueName: \"kubernetes.io/projected/04de4a38-5d09-4c68-b7a2-be6482748e50-kube-api-access-8rzbx\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.537352 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04de4a38-5d09-4c68-b7a2-be6482748e50-logs\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.538173 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebcbc04a-230e-46b1-be0d-af9ece769369-kube-api-access-ls4wl" (OuterVolumeSpecName: "kube-api-access-ls4wl") pod "ebcbc04a-230e-46b1-be0d-af9ece769369" (UID: "ebcbc04a-230e-46b1-be0d-af9ece769369"). InnerVolumeSpecName "kube-api-access-ls4wl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.541325 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ebcbc04a-230e-46b1-be0d-af9ece769369-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "ebcbc04a-230e-46b1-be0d-af9ece769369" (UID: "ebcbc04a-230e-46b1-be0d-af9ece769369"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.582876 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ebcbc04a-230e-46b1-be0d-af9ece769369-config-data" (OuterVolumeSpecName: "config-data") pod "ebcbc04a-230e-46b1-be0d-af9ece769369" (UID: "ebcbc04a-230e-46b1-be0d-af9ece769369"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.596356 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.608582 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ebcbc04a-230e-46b1-be0d-af9ece769369-scripts" (OuterVolumeSpecName: "scripts") pod "ebcbc04a-230e-46b1-be0d-af9ece769369" (UID: "ebcbc04a-230e-46b1-be0d-af9ece769369"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.629135 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.639004 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e378b38-1cb5-4f77-b9a4-e6cfce8da188-combined-ca-bundle\") pod \"9e378b38-1cb5-4f77-b9a4-e6cfce8da188\" (UID: \"9e378b38-1cb5-4f77-b9a4-e6cfce8da188\") " Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.639087 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e378b38-1cb5-4f77-b9a4-e6cfce8da188-ceilometer-tls-certs\") pod \"9e378b38-1cb5-4f77-b9a4-e6cfce8da188\" (UID: \"9e378b38-1cb5-4f77-b9a4-e6cfce8da188\") " Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.639160 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9e378b38-1cb5-4f77-b9a4-e6cfce8da188-sg-core-conf-yaml\") pod \"9e378b38-1cb5-4f77-b9a4-e6cfce8da188\" (UID: \"9e378b38-1cb5-4f77-b9a4-e6cfce8da188\") " Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.639237 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e378b38-1cb5-4f77-b9a4-e6cfce8da188-scripts\") pod \"9e378b38-1cb5-4f77-b9a4-e6cfce8da188\" (UID: \"9e378b38-1cb5-4f77-b9a4-e6cfce8da188\") " Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.639354 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e378b38-1cb5-4f77-b9a4-e6cfce8da188-config-data\") pod \"9e378b38-1cb5-4f77-b9a4-e6cfce8da188\" (UID: \"9e378b38-1cb5-4f77-b9a4-e6cfce8da188\") " Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.639551 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9cfg5\" (UniqueName: \"kubernetes.io/projected/9e378b38-1cb5-4f77-b9a4-e6cfce8da188-kube-api-access-9cfg5\") pod \"9e378b38-1cb5-4f77-b9a4-e6cfce8da188\" (UID: \"9e378b38-1cb5-4f77-b9a4-e6cfce8da188\") " Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.639591 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9e378b38-1cb5-4f77-b9a4-e6cfce8da188-log-httpd\") pod \"9e378b38-1cb5-4f77-b9a4-e6cfce8da188\" (UID: \"9e378b38-1cb5-4f77-b9a4-e6cfce8da188\") " Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.639696 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9e378b38-1cb5-4f77-b9a4-e6cfce8da188-run-httpd\") pod \"9e378b38-1cb5-4f77-b9a4-e6cfce8da188\" (UID: \"9e378b38-1cb5-4f77-b9a4-e6cfce8da188\") " Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.641410 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ebcbc04a-230e-46b1-be0d-af9ece769369-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.641436 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ls4wl\" (UniqueName: \"kubernetes.io/projected/ebcbc04a-230e-46b1-be0d-af9ece769369-kube-api-access-ls4wl\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.641452 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ebcbc04a-230e-46b1-be0d-af9ece769369-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.641466 4799 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ebcbc04a-230e-46b1-be0d-af9ece769369-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.642677 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e378b38-1cb5-4f77-b9a4-e6cfce8da188-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "9e378b38-1cb5-4f77-b9a4-e6cfce8da188" (UID: "9e378b38-1cb5-4f77-b9a4-e6cfce8da188"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.654841 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e378b38-1cb5-4f77-b9a4-e6cfce8da188-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "9e378b38-1cb5-4f77-b9a4-e6cfce8da188" (UID: "9e378b38-1cb5-4f77-b9a4-e6cfce8da188"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.660766 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e378b38-1cb5-4f77-b9a4-e6cfce8da188-scripts" (OuterVolumeSpecName: "scripts") pod "9e378b38-1cb5-4f77-b9a4-e6cfce8da188" (UID: "9e378b38-1cb5-4f77-b9a4-e6cfce8da188"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.667067 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e378b38-1cb5-4f77-b9a4-e6cfce8da188-kube-api-access-9cfg5" (OuterVolumeSpecName: "kube-api-access-9cfg5") pod "9e378b38-1cb5-4f77-b9a4-e6cfce8da188" (UID: "9e378b38-1cb5-4f77-b9a4-e6cfce8da188"). InnerVolumeSpecName "kube-api-access-9cfg5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.700352 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e378b38-1cb5-4f77-b9a4-e6cfce8da188-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "9e378b38-1cb5-4f77-b9a4-e6cfce8da188" (UID: "9e378b38-1cb5-4f77-b9a4-e6cfce8da188"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.700998 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.728608 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e378b38-1cb5-4f77-b9a4-e6cfce8da188-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "9e378b38-1cb5-4f77-b9a4-e6cfce8da188" (UID: "9e378b38-1cb5-4f77-b9a4-e6cfce8da188"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.749741 4799 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9e378b38-1cb5-4f77-b9a4-e6cfce8da188-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.749770 4799 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e378b38-1cb5-4f77-b9a4-e6cfce8da188-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.749782 4799 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9e378b38-1cb5-4f77-b9a4-e6cfce8da188-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.749803 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e378b38-1cb5-4f77-b9a4-e6cfce8da188-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.749814 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9cfg5\" (UniqueName: \"kubernetes.io/projected/9e378b38-1cb5-4f77-b9a4-e6cfce8da188-kube-api-access-9cfg5\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.749825 4799 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9e378b38-1cb5-4f77-b9a4-e6cfce8da188-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.815529 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e378b38-1cb5-4f77-b9a4-e6cfce8da188-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9e378b38-1cb5-4f77-b9a4-e6cfce8da188" (UID: "9e378b38-1cb5-4f77-b9a4-e6cfce8da188"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.840610 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e378b38-1cb5-4f77-b9a4-e6cfce8da188-config-data" (OuterVolumeSpecName: "config-data") pod "9e378b38-1cb5-4f77-b9a4-e6cfce8da188" (UID: "9e378b38-1cb5-4f77-b9a4-e6cfce8da188"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.852257 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e378b38-1cb5-4f77-b9a4-e6cfce8da188-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.852295 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e378b38-1cb5-4f77-b9a4-e6cfce8da188-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.873040 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-76b5fdb995-cbg8h" Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.970855 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-864d5fc68c-cws22"] Nov 29 05:29:30 crc kubenswrapper[4799]: I1129 05:29:30.971124 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-864d5fc68c-cws22" podUID="ee60cfe8-e5d9-4e9a-9299-38a14710fcdd" containerName="dnsmasq-dns" containerID="cri-o://c51c2754d228077a09b73dce28006df605dc6d539e565145ce5b6edffcfe0de8" gracePeriod=10 Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.098453 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-7c5488f496-jvw7p" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.193879 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-86d9b546cb-5xxhb"] Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.194137 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-86d9b546cb-5xxhb" podUID="f09087c5-87d5-4cd1-9a91-cbe48713d9e9" containerName="horizon-log" containerID="cri-o://8f1f666f1cb9af792d04255e0b4e5900be38a9d0b4060fd4c97f314b6c12a164" gracePeriod=30 Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.195581 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-86d9b546cb-5xxhb" podUID="f09087c5-87d5-4cd1-9a91-cbe48713d9e9" containerName="horizon" containerID="cri-o://acf62f8dda4e17622de5c459a6b4fe1c70a55e3481d77ef29c8a3bc868f9df26" gracePeriod=30 Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.215038 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-86d9b546cb-5xxhb" podUID="f09087c5-87d5-4cd1-9a91-cbe48713d9e9" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.236:8443/dashboard/auth/login/?next=/dashboard/\": EOF" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.235523 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-566ff59745-jp249" event={"ID":"ebcbc04a-230e-46b1-be0d-af9ece769369","Type":"ContainerDied","Data":"b7b41b27a0067adde7d22f2cc2d1916b9a2cfbbfcc3eeb4732115a1442f63e6a"} Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.235595 4799 scope.go:117] "RemoveContainer" containerID="72186f35f25c7e3357a6a3815736e28ff75be3b34cc5516c50542e4d500121dc" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.235781 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-566ff59745-jp249" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.243611 4799 generic.go:334] "Generic (PLEG): container finished" podID="ee60cfe8-e5d9-4e9a-9299-38a14710fcdd" containerID="c51c2754d228077a09b73dce28006df605dc6d539e565145ce5b6edffcfe0de8" exitCode=0 Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.243664 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-864d5fc68c-cws22" event={"ID":"ee60cfe8-e5d9-4e9a-9299-38a14710fcdd","Type":"ContainerDied","Data":"c51c2754d228077a09b73dce28006df605dc6d539e565145ce5b6edffcfe0de8"} Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.246521 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"73c23628-dfd6-4496-9091-c1fbe3b092f6","Type":"ContainerStarted","Data":"bac0acaab63c16ad1ff009ee59546948c92e8b23343fb2825841b80f44e49f08"} Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.254206 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9e378b38-1cb5-4f77-b9a4-e6cfce8da188","Type":"ContainerDied","Data":"f72ac746d459003c055d5bd70bdc1d61bb792cd8aa649c5218ef2ca22d8f7cf7"} Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.254292 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.261986 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-598444b44c-v5znr" event={"ID":"04de4a38-5d09-4c68-b7a2-be6482748e50","Type":"ContainerDied","Data":"af552d396b39049c076f4bbbd89611106fca62ea40b2ec616195df247cba534e"} Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.262158 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-598444b44c-v5znr" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.265835 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"b8b16394-158d-47e0-b406-18636e57be57","Type":"ContainerStarted","Data":"0aeb6766a87b7e403cb49bca82960f4cf7d3deab883f6e6d87601744c60dec53"} Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.307987 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-566ff59745-jp249"] Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.319451 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-566ff59745-jp249"] Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.344951 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.364802 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.384268 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-598444b44c-v5znr"] Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.404997 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-598444b44c-v5znr"] Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.422126 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 29 05:29:31 crc kubenswrapper[4799]: E1129 05:29:31.424093 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e378b38-1cb5-4f77-b9a4-e6cfce8da188" containerName="ceilometer-notification-agent" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.424209 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e378b38-1cb5-4f77-b9a4-e6cfce8da188" containerName="ceilometer-notification-agent" Nov 29 05:29:31 crc kubenswrapper[4799]: E1129 05:29:31.424243 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebcbc04a-230e-46b1-be0d-af9ece769369" containerName="horizon" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.424250 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebcbc04a-230e-46b1-be0d-af9ece769369" containerName="horizon" Nov 29 05:29:31 crc kubenswrapper[4799]: E1129 05:29:31.424327 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e378b38-1cb5-4f77-b9a4-e6cfce8da188" containerName="sg-core" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.424360 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e378b38-1cb5-4f77-b9a4-e6cfce8da188" containerName="sg-core" Nov 29 05:29:31 crc kubenswrapper[4799]: E1129 05:29:31.424414 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04de4a38-5d09-4c68-b7a2-be6482748e50" containerName="horizon" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.424447 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="04de4a38-5d09-4c68-b7a2-be6482748e50" containerName="horizon" Nov 29 05:29:31 crc kubenswrapper[4799]: E1129 05:29:31.424466 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e378b38-1cb5-4f77-b9a4-e6cfce8da188" containerName="ceilometer-central-agent" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.424474 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e378b38-1cb5-4f77-b9a4-e6cfce8da188" containerName="ceilometer-central-agent" Nov 29 05:29:31 crc kubenswrapper[4799]: E1129 05:29:31.424495 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e378b38-1cb5-4f77-b9a4-e6cfce8da188" containerName="proxy-httpd" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.424504 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e378b38-1cb5-4f77-b9a4-e6cfce8da188" containerName="proxy-httpd" Nov 29 05:29:31 crc kubenswrapper[4799]: E1129 05:29:31.424552 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebcbc04a-230e-46b1-be0d-af9ece769369" containerName="horizon-log" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.424560 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebcbc04a-230e-46b1-be0d-af9ece769369" containerName="horizon-log" Nov 29 05:29:31 crc kubenswrapper[4799]: E1129 05:29:31.424569 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04de4a38-5d09-4c68-b7a2-be6482748e50" containerName="horizon-log" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.424577 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="04de4a38-5d09-4c68-b7a2-be6482748e50" containerName="horizon-log" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.425095 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e378b38-1cb5-4f77-b9a4-e6cfce8da188" containerName="sg-core" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.425123 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebcbc04a-230e-46b1-be0d-af9ece769369" containerName="horizon-log" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.425140 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e378b38-1cb5-4f77-b9a4-e6cfce8da188" containerName="ceilometer-notification-agent" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.425155 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e378b38-1cb5-4f77-b9a4-e6cfce8da188" containerName="proxy-httpd" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.425169 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="04de4a38-5d09-4c68-b7a2-be6482748e50" containerName="horizon-log" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.425179 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e378b38-1cb5-4f77-b9a4-e6cfce8da188" containerName="ceilometer-central-agent" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.425190 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="04de4a38-5d09-4c68-b7a2-be6482748e50" containerName="horizon" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.425200 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebcbc04a-230e-46b1-be0d-af9ece769369" containerName="horizon" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.427544 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.431477 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.434332 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.436374 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.464279 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.472177 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a9740510-775e-4907-8591-40f7a97e10e3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a9740510-775e-4907-8591-40f7a97e10e3\") " pod="openstack/ceilometer-0" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.472261 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9740510-775e-4907-8591-40f7a97e10e3-config-data\") pod \"ceilometer-0\" (UID: \"a9740510-775e-4907-8591-40f7a97e10e3\") " pod="openstack/ceilometer-0" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.472381 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9740510-775e-4907-8591-40f7a97e10e3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a9740510-775e-4907-8591-40f7a97e10e3\") " pod="openstack/ceilometer-0" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.472414 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9740510-775e-4907-8591-40f7a97e10e3-run-httpd\") pod \"ceilometer-0\" (UID: \"a9740510-775e-4907-8591-40f7a97e10e3\") " pod="openstack/ceilometer-0" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.473379 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9740510-775e-4907-8591-40f7a97e10e3-log-httpd\") pod \"ceilometer-0\" (UID: \"a9740510-775e-4907-8591-40f7a97e10e3\") " pod="openstack/ceilometer-0" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.473410 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9740510-775e-4907-8591-40f7a97e10e3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a9740510-775e-4907-8591-40f7a97e10e3\") " pod="openstack/ceilometer-0" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.473498 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49tfp\" (UniqueName: \"kubernetes.io/projected/a9740510-775e-4907-8591-40f7a97e10e3-kube-api-access-49tfp\") pod \"ceilometer-0\" (UID: \"a9740510-775e-4907-8591-40f7a97e10e3\") " pod="openstack/ceilometer-0" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.473589 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9740510-775e-4907-8591-40f7a97e10e3-scripts\") pod \"ceilometer-0\" (UID: \"a9740510-775e-4907-8591-40f7a97e10e3\") " pod="openstack/ceilometer-0" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.527271 4799 scope.go:117] "RemoveContainer" containerID="a8ed286597384b45acd04d7e67e5c08d350f935fef4e4fbb30c795f46c0acf50" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.576980 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49tfp\" (UniqueName: \"kubernetes.io/projected/a9740510-775e-4907-8591-40f7a97e10e3-kube-api-access-49tfp\") pod \"ceilometer-0\" (UID: \"a9740510-775e-4907-8591-40f7a97e10e3\") " pod="openstack/ceilometer-0" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.577059 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9740510-775e-4907-8591-40f7a97e10e3-scripts\") pod \"ceilometer-0\" (UID: \"a9740510-775e-4907-8591-40f7a97e10e3\") " pod="openstack/ceilometer-0" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.577138 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a9740510-775e-4907-8591-40f7a97e10e3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a9740510-775e-4907-8591-40f7a97e10e3\") " pod="openstack/ceilometer-0" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.577161 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9740510-775e-4907-8591-40f7a97e10e3-config-data\") pod \"ceilometer-0\" (UID: \"a9740510-775e-4907-8591-40f7a97e10e3\") " pod="openstack/ceilometer-0" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.577203 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9740510-775e-4907-8591-40f7a97e10e3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a9740510-775e-4907-8591-40f7a97e10e3\") " pod="openstack/ceilometer-0" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.577228 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9740510-775e-4907-8591-40f7a97e10e3-run-httpd\") pod \"ceilometer-0\" (UID: \"a9740510-775e-4907-8591-40f7a97e10e3\") " pod="openstack/ceilometer-0" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.577295 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9740510-775e-4907-8591-40f7a97e10e3-log-httpd\") pod \"ceilometer-0\" (UID: \"a9740510-775e-4907-8591-40f7a97e10e3\") " pod="openstack/ceilometer-0" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.577312 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9740510-775e-4907-8591-40f7a97e10e3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a9740510-775e-4907-8591-40f7a97e10e3\") " pod="openstack/ceilometer-0" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.584076 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9740510-775e-4907-8591-40f7a97e10e3-run-httpd\") pod \"ceilometer-0\" (UID: \"a9740510-775e-4907-8591-40f7a97e10e3\") " pod="openstack/ceilometer-0" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.585755 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9740510-775e-4907-8591-40f7a97e10e3-config-data\") pod \"ceilometer-0\" (UID: \"a9740510-775e-4907-8591-40f7a97e10e3\") " pod="openstack/ceilometer-0" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.587777 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9740510-775e-4907-8591-40f7a97e10e3-log-httpd\") pod \"ceilometer-0\" (UID: \"a9740510-775e-4907-8591-40f7a97e10e3\") " pod="openstack/ceilometer-0" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.596447 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a9740510-775e-4907-8591-40f7a97e10e3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a9740510-775e-4907-8591-40f7a97e10e3\") " pod="openstack/ceilometer-0" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.596784 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9740510-775e-4907-8591-40f7a97e10e3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a9740510-775e-4907-8591-40f7a97e10e3\") " pod="openstack/ceilometer-0" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.617369 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9740510-775e-4907-8591-40f7a97e10e3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a9740510-775e-4907-8591-40f7a97e10e3\") " pod="openstack/ceilometer-0" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.617645 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9740510-775e-4907-8591-40f7a97e10e3-scripts\") pod \"ceilometer-0\" (UID: \"a9740510-775e-4907-8591-40f7a97e10e3\") " pod="openstack/ceilometer-0" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.624386 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49tfp\" (UniqueName: \"kubernetes.io/projected/a9740510-775e-4907-8591-40f7a97e10e3-kube-api-access-49tfp\") pod \"ceilometer-0\" (UID: \"a9740510-775e-4907-8591-40f7a97e10e3\") " pod="openstack/ceilometer-0" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.724451 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-864d5fc68c-cws22" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.758376 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.776918 4799 scope.go:117] "RemoveContainer" containerID="be4294eae32e4ed3aef4ebf11a13fbb8fb8da2c722a9d820a58674f8e9a082dc" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.793240 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ee60cfe8-e5d9-4e9a-9299-38a14710fcdd-ovsdbserver-nb\") pod \"ee60cfe8-e5d9-4e9a-9299-38a14710fcdd\" (UID: \"ee60cfe8-e5d9-4e9a-9299-38a14710fcdd\") " Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.793322 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ee60cfe8-e5d9-4e9a-9299-38a14710fcdd-ovsdbserver-sb\") pod \"ee60cfe8-e5d9-4e9a-9299-38a14710fcdd\" (UID: \"ee60cfe8-e5d9-4e9a-9299-38a14710fcdd\") " Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.793397 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee60cfe8-e5d9-4e9a-9299-38a14710fcdd-dns-svc\") pod \"ee60cfe8-e5d9-4e9a-9299-38a14710fcdd\" (UID: \"ee60cfe8-e5d9-4e9a-9299-38a14710fcdd\") " Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.793448 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-87595\" (UniqueName: \"kubernetes.io/projected/ee60cfe8-e5d9-4e9a-9299-38a14710fcdd-kube-api-access-87595\") pod \"ee60cfe8-e5d9-4e9a-9299-38a14710fcdd\" (UID: \"ee60cfe8-e5d9-4e9a-9299-38a14710fcdd\") " Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.793573 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee60cfe8-e5d9-4e9a-9299-38a14710fcdd-config\") pod \"ee60cfe8-e5d9-4e9a-9299-38a14710fcdd\" (UID: \"ee60cfe8-e5d9-4e9a-9299-38a14710fcdd\") " Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.793631 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ee60cfe8-e5d9-4e9a-9299-38a14710fcdd-openstack-edpm-ipam\") pod \"ee60cfe8-e5d9-4e9a-9299-38a14710fcdd\" (UID: \"ee60cfe8-e5d9-4e9a-9299-38a14710fcdd\") " Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.803948 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee60cfe8-e5d9-4e9a-9299-38a14710fcdd-kube-api-access-87595" (OuterVolumeSpecName: "kube-api-access-87595") pod "ee60cfe8-e5d9-4e9a-9299-38a14710fcdd" (UID: "ee60cfe8-e5d9-4e9a-9299-38a14710fcdd"). InnerVolumeSpecName "kube-api-access-87595". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.821544 4799 scope.go:117] "RemoveContainer" containerID="adc14bfee676568c6a1b21f35b94c262a6fe0699e51bf4453d9d71b62855b2d8" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.872245 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee60cfe8-e5d9-4e9a-9299-38a14710fcdd-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "ee60cfe8-e5d9-4e9a-9299-38a14710fcdd" (UID: "ee60cfe8-e5d9-4e9a-9299-38a14710fcdd"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.877584 4799 scope.go:117] "RemoveContainer" containerID="d09b5086c0cd8c7c5bb69705d115dec19ed06948186960f1aa1eafe1b7cd88a8" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.882632 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee60cfe8-e5d9-4e9a-9299-38a14710fcdd-config" (OuterVolumeSpecName: "config") pod "ee60cfe8-e5d9-4e9a-9299-38a14710fcdd" (UID: "ee60cfe8-e5d9-4e9a-9299-38a14710fcdd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.882904 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee60cfe8-e5d9-4e9a-9299-38a14710fcdd-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ee60cfe8-e5d9-4e9a-9299-38a14710fcdd" (UID: "ee60cfe8-e5d9-4e9a-9299-38a14710fcdd"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.901568 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee60cfe8-e5d9-4e9a-9299-38a14710fcdd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ee60cfe8-e5d9-4e9a-9299-38a14710fcdd" (UID: "ee60cfe8-e5d9-4e9a-9299-38a14710fcdd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.903514 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ee60cfe8-e5d9-4e9a-9299-38a14710fcdd-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.903573 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee60cfe8-e5d9-4e9a-9299-38a14710fcdd-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.903585 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-87595\" (UniqueName: \"kubernetes.io/projected/ee60cfe8-e5d9-4e9a-9299-38a14710fcdd-kube-api-access-87595\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.903598 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee60cfe8-e5d9-4e9a-9299-38a14710fcdd-config\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.903609 4799 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ee60cfe8-e5d9-4e9a-9299-38a14710fcdd-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:31 crc kubenswrapper[4799]: I1129 05:29:31.908051 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee60cfe8-e5d9-4e9a-9299-38a14710fcdd-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ee60cfe8-e5d9-4e9a-9299-38a14710fcdd" (UID: "ee60cfe8-e5d9-4e9a-9299-38a14710fcdd"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 05:29:32 crc kubenswrapper[4799]: I1129 05:29:32.008341 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ee60cfe8-e5d9-4e9a-9299-38a14710fcdd-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:32 crc kubenswrapper[4799]: I1129 05:29:32.036483 4799 scope.go:117] "RemoveContainer" containerID="86e09030d07ac79226803571fd562d1b65201ff6597388493a2af883ba02cab8" Nov 29 05:29:32 crc kubenswrapper[4799]: I1129 05:29:32.089103 4799 scope.go:117] "RemoveContainer" containerID="f8801fb2760fa321e80c168af0997ee85b677223a66f5991257e83509f823137" Nov 29 05:29:32 crc kubenswrapper[4799]: I1129 05:29:32.260317 4799 scope.go:117] "RemoveContainer" containerID="69cac5b4704d7a992b979fd8a84a6d26615f488bf691883048255f969711a221" Nov 29 05:29:32 crc kubenswrapper[4799]: I1129 05:29:32.284047 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"73c23628-dfd6-4496-9091-c1fbe3b092f6","Type":"ContainerStarted","Data":"7636518fd1975296dddcdd024b7b7c7f8de8fa7adcb8789a19fe34bbcc89eb50"} Nov 29 05:29:32 crc kubenswrapper[4799]: I1129 05:29:32.295011 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"b8b16394-158d-47e0-b406-18636e57be57","Type":"ContainerStarted","Data":"3974d3b416db3e05d6508850c3371295124f2fb92eb30d60fa945fcb43e59e54"} Nov 29 05:29:32 crc kubenswrapper[4799]: I1129 05:29:32.310537 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-864d5fc68c-cws22" event={"ID":"ee60cfe8-e5d9-4e9a-9299-38a14710fcdd","Type":"ContainerDied","Data":"8938a1302f500518ad2f3852e6fdc2286b8b29860d55d2dc2bc6c920623bdf5d"} Nov 29 05:29:32 crc kubenswrapper[4799]: I1129 05:29:32.310615 4799 scope.go:117] "RemoveContainer" containerID="c51c2754d228077a09b73dce28006df605dc6d539e565145ce5b6edffcfe0de8" Nov 29 05:29:32 crc kubenswrapper[4799]: I1129 05:29:32.310827 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-864d5fc68c-cws22" Nov 29 05:29:32 crc kubenswrapper[4799]: I1129 05:29:32.315666 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=3.902472919 podStartE2EDuration="12.31563866s" podCreationTimestamp="2025-11-29 05:29:20 +0000 UTC" firstStartedPulling="2025-11-29 05:29:21.572786957 +0000 UTC m=+3037.215717377" lastFinishedPulling="2025-11-29 05:29:29.985952718 +0000 UTC m=+3045.628883118" observedRunningTime="2025-11-29 05:29:32.307902966 +0000 UTC m=+3047.950833376" watchObservedRunningTime="2025-11-29 05:29:32.31563866 +0000 UTC m=+3047.958569060" Nov 29 05:29:32 crc kubenswrapper[4799]: I1129 05:29:32.338470 4799 scope.go:117] "RemoveContainer" containerID="f766d10c1e89373de58494f3c5230999c5644629c91ff54a32ee511452b165c5" Nov 29 05:29:32 crc kubenswrapper[4799]: I1129 05:29:32.423584 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-864d5fc68c-cws22"] Nov 29 05:29:32 crc kubenswrapper[4799]: I1129 05:29:32.428044 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 29 05:29:32 crc kubenswrapper[4799]: I1129 05:29:32.437497 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-864d5fc68c-cws22"] Nov 29 05:29:32 crc kubenswrapper[4799]: E1129 05:29:32.498511 4799 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee60cfe8_e5d9_4e9a_9299_38a14710fcdd.slice\": RecentStats: unable to find data in memory cache]" Nov 29 05:29:32 crc kubenswrapper[4799]: I1129 05:29:32.671197 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04de4a38-5d09-4c68-b7a2-be6482748e50" path="/var/lib/kubelet/pods/04de4a38-5d09-4c68-b7a2-be6482748e50/volumes" Nov 29 05:29:32 crc kubenswrapper[4799]: I1129 05:29:32.672540 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e378b38-1cb5-4f77-b9a4-e6cfce8da188" path="/var/lib/kubelet/pods/9e378b38-1cb5-4f77-b9a4-e6cfce8da188/volumes" Nov 29 05:29:32 crc kubenswrapper[4799]: I1129 05:29:32.674027 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ebcbc04a-230e-46b1-be0d-af9ece769369" path="/var/lib/kubelet/pods/ebcbc04a-230e-46b1-be0d-af9ece769369/volumes" Nov 29 05:29:32 crc kubenswrapper[4799]: I1129 05:29:32.674594 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee60cfe8-e5d9-4e9a-9299-38a14710fcdd" path="/var/lib/kubelet/pods/ee60cfe8-e5d9-4e9a-9299-38a14710fcdd/volumes" Nov 29 05:29:33 crc kubenswrapper[4799]: I1129 05:29:33.351980 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"b8b16394-158d-47e0-b406-18636e57be57","Type":"ContainerStarted","Data":"6fa80286f341732c424d74b791793e9868a1a88efd2e95ea536e61f0614a1531"} Nov 29 05:29:33 crc kubenswrapper[4799]: I1129 05:29:33.352464 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Nov 29 05:29:33 crc kubenswrapper[4799]: I1129 05:29:33.364083 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9740510-775e-4907-8591-40f7a97e10e3","Type":"ContainerStarted","Data":"762dacdd1ed6d04400f83778da5bc493378e06346e9026fad12bc1998c668365"} Nov 29 05:29:33 crc kubenswrapper[4799]: I1129 05:29:33.364137 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9740510-775e-4907-8591-40f7a97e10e3","Type":"ContainerStarted","Data":"8ee07f134c04fa49d98990799f60a86f73eb656b01232b161cf1757f8f1b6f17"} Nov 29 05:29:33 crc kubenswrapper[4799]: I1129 05:29:33.395442 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=6.395411177 podStartE2EDuration="6.395411177s" podCreationTimestamp="2025-11-29 05:29:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 05:29:33.37967271 +0000 UTC m=+3049.022603110" watchObservedRunningTime="2025-11-29 05:29:33.395411177 +0000 UTC m=+3049.038341577" Nov 29 05:29:34 crc kubenswrapper[4799]: I1129 05:29:34.313980 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 29 05:29:34 crc kubenswrapper[4799]: I1129 05:29:34.352605 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-86d9b546cb-5xxhb" podUID="f09087c5-87d5-4cd1-9a91-cbe48713d9e9" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.236:8443/dashboard/auth/login/?next=/dashboard/\": read tcp 10.217.0.2:44054->10.217.0.236:8443: read: connection reset by peer" Nov 29 05:29:34 crc kubenswrapper[4799]: I1129 05:29:34.382568 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9740510-775e-4907-8591-40f7a97e10e3","Type":"ContainerStarted","Data":"2110e05ff74ea55f482c4349e2706399889e31bbc793d354adcb30f970f95061"} Nov 29 05:29:34 crc kubenswrapper[4799]: I1129 05:29:34.671019 4799 scope.go:117] "RemoveContainer" containerID="e6d4b3d1d5b5e1eeb13508d67de3d3c3aac52e76337b55e53793185c0d001d82" Nov 29 05:29:34 crc kubenswrapper[4799]: E1129 05:29:34.671331 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:29:35 crc kubenswrapper[4799]: I1129 05:29:35.402475 4799 generic.go:334] "Generic (PLEG): container finished" podID="f09087c5-87d5-4cd1-9a91-cbe48713d9e9" containerID="acf62f8dda4e17622de5c459a6b4fe1c70a55e3481d77ef29c8a3bc868f9df26" exitCode=0 Nov 29 05:29:35 crc kubenswrapper[4799]: I1129 05:29:35.402562 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-86d9b546cb-5xxhb" event={"ID":"f09087c5-87d5-4cd1-9a91-cbe48713d9e9","Type":"ContainerDied","Data":"acf62f8dda4e17622de5c459a6b4fe1c70a55e3481d77ef29c8a3bc868f9df26"} Nov 29 05:29:35 crc kubenswrapper[4799]: I1129 05:29:35.407573 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9740510-775e-4907-8591-40f7a97e10e3","Type":"ContainerStarted","Data":"0ebd7759b073d3dc35b43aae0e7f2beb369aa26b14ae4883b9f5ceb0a0451a60"} Nov 29 05:29:36 crc kubenswrapper[4799]: I1129 05:29:36.799601 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-86d9b546cb-5xxhb" podUID="f09087c5-87d5-4cd1-9a91-cbe48713d9e9" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.236:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.236:8443: connect: connection refused" Nov 29 05:29:37 crc kubenswrapper[4799]: I1129 05:29:37.430988 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9740510-775e-4907-8591-40f7a97e10e3","Type":"ContainerStarted","Data":"ec80651a1e3eec9a4db39805031ea01ef9f62fda7f6b36da3a77b052d1ec07da"} Nov 29 05:29:37 crc kubenswrapper[4799]: I1129 05:29:37.431434 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 29 05:29:37 crc kubenswrapper[4799]: I1129 05:29:37.431418 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a9740510-775e-4907-8591-40f7a97e10e3" containerName="proxy-httpd" containerID="cri-o://ec80651a1e3eec9a4db39805031ea01ef9f62fda7f6b36da3a77b052d1ec07da" gracePeriod=30 Nov 29 05:29:37 crc kubenswrapper[4799]: I1129 05:29:37.431716 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a9740510-775e-4907-8591-40f7a97e10e3" containerName="sg-core" containerID="cri-o://0ebd7759b073d3dc35b43aae0e7f2beb369aa26b14ae4883b9f5ceb0a0451a60" gracePeriod=30 Nov 29 05:29:37 crc kubenswrapper[4799]: I1129 05:29:37.431711 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a9740510-775e-4907-8591-40f7a97e10e3" containerName="ceilometer-notification-agent" containerID="cri-o://2110e05ff74ea55f482c4349e2706399889e31bbc793d354adcb30f970f95061" gracePeriod=30 Nov 29 05:29:37 crc kubenswrapper[4799]: I1129 05:29:37.431851 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a9740510-775e-4907-8591-40f7a97e10e3" containerName="ceilometer-central-agent" containerID="cri-o://762dacdd1ed6d04400f83778da5bc493378e06346e9026fad12bc1998c668365" gracePeriod=30 Nov 29 05:29:37 crc kubenswrapper[4799]: I1129 05:29:37.473031 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.906608152 podStartE2EDuration="6.473004739s" podCreationTimestamp="2025-11-29 05:29:31 +0000 UTC" firstStartedPulling="2025-11-29 05:29:32.39652057 +0000 UTC m=+3048.039450970" lastFinishedPulling="2025-11-29 05:29:36.962917127 +0000 UTC m=+3052.605847557" observedRunningTime="2025-11-29 05:29:37.465298004 +0000 UTC m=+3053.108228404" watchObservedRunningTime="2025-11-29 05:29:37.473004739 +0000 UTC m=+3053.115935149" Nov 29 05:29:38 crc kubenswrapper[4799]: I1129 05:29:38.462446 4799 generic.go:334] "Generic (PLEG): container finished" podID="a9740510-775e-4907-8591-40f7a97e10e3" containerID="ec80651a1e3eec9a4db39805031ea01ef9f62fda7f6b36da3a77b052d1ec07da" exitCode=0 Nov 29 05:29:38 crc kubenswrapper[4799]: I1129 05:29:38.463731 4799 generic.go:334] "Generic (PLEG): container finished" podID="a9740510-775e-4907-8591-40f7a97e10e3" containerID="0ebd7759b073d3dc35b43aae0e7f2beb369aa26b14ae4883b9f5ceb0a0451a60" exitCode=2 Nov 29 05:29:38 crc kubenswrapper[4799]: I1129 05:29:38.462536 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9740510-775e-4907-8591-40f7a97e10e3","Type":"ContainerDied","Data":"ec80651a1e3eec9a4db39805031ea01ef9f62fda7f6b36da3a77b052d1ec07da"} Nov 29 05:29:38 crc kubenswrapper[4799]: I1129 05:29:38.463937 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9740510-775e-4907-8591-40f7a97e10e3","Type":"ContainerDied","Data":"0ebd7759b073d3dc35b43aae0e7f2beb369aa26b14ae4883b9f5ceb0a0451a60"} Nov 29 05:29:38 crc kubenswrapper[4799]: I1129 05:29:38.463864 4799 generic.go:334] "Generic (PLEG): container finished" podID="a9740510-775e-4907-8591-40f7a97e10e3" containerID="2110e05ff74ea55f482c4349e2706399889e31bbc793d354adcb30f970f95061" exitCode=0 Nov 29 05:29:38 crc kubenswrapper[4799]: I1129 05:29:38.463969 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9740510-775e-4907-8591-40f7a97e10e3","Type":"ContainerDied","Data":"2110e05ff74ea55f482c4349e2706399889e31bbc793d354adcb30f970f95061"} Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.318266 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.414845 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9740510-775e-4907-8591-40f7a97e10e3-scripts\") pod \"a9740510-775e-4907-8591-40f7a97e10e3\" (UID: \"a9740510-775e-4907-8591-40f7a97e10e3\") " Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.414981 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9740510-775e-4907-8591-40f7a97e10e3-run-httpd\") pod \"a9740510-775e-4907-8591-40f7a97e10e3\" (UID: \"a9740510-775e-4907-8591-40f7a97e10e3\") " Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.415029 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9740510-775e-4907-8591-40f7a97e10e3-config-data\") pod \"a9740510-775e-4907-8591-40f7a97e10e3\" (UID: \"a9740510-775e-4907-8591-40f7a97e10e3\") " Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.415187 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9740510-775e-4907-8591-40f7a97e10e3-combined-ca-bundle\") pod \"a9740510-775e-4907-8591-40f7a97e10e3\" (UID: \"a9740510-775e-4907-8591-40f7a97e10e3\") " Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.415236 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a9740510-775e-4907-8591-40f7a97e10e3-sg-core-conf-yaml\") pod \"a9740510-775e-4907-8591-40f7a97e10e3\" (UID: \"a9740510-775e-4907-8591-40f7a97e10e3\") " Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.415327 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-49tfp\" (UniqueName: \"kubernetes.io/projected/a9740510-775e-4907-8591-40f7a97e10e3-kube-api-access-49tfp\") pod \"a9740510-775e-4907-8591-40f7a97e10e3\" (UID: \"a9740510-775e-4907-8591-40f7a97e10e3\") " Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.415369 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9740510-775e-4907-8591-40f7a97e10e3-ceilometer-tls-certs\") pod \"a9740510-775e-4907-8591-40f7a97e10e3\" (UID: \"a9740510-775e-4907-8591-40f7a97e10e3\") " Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.415388 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9740510-775e-4907-8591-40f7a97e10e3-log-httpd\") pod \"a9740510-775e-4907-8591-40f7a97e10e3\" (UID: \"a9740510-775e-4907-8591-40f7a97e10e3\") " Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.416764 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9740510-775e-4907-8591-40f7a97e10e3-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a9740510-775e-4907-8591-40f7a97e10e3" (UID: "a9740510-775e-4907-8591-40f7a97e10e3"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.417762 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9740510-775e-4907-8591-40f7a97e10e3-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a9740510-775e-4907-8591-40f7a97e10e3" (UID: "a9740510-775e-4907-8591-40f7a97e10e3"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.423054 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9740510-775e-4907-8591-40f7a97e10e3-kube-api-access-49tfp" (OuterVolumeSpecName: "kube-api-access-49tfp") pod "a9740510-775e-4907-8591-40f7a97e10e3" (UID: "a9740510-775e-4907-8591-40f7a97e10e3"). InnerVolumeSpecName "kube-api-access-49tfp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.423756 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9740510-775e-4907-8591-40f7a97e10e3-scripts" (OuterVolumeSpecName: "scripts") pod "a9740510-775e-4907-8591-40f7a97e10e3" (UID: "a9740510-775e-4907-8591-40f7a97e10e3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.479155 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9740510-775e-4907-8591-40f7a97e10e3-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a9740510-775e-4907-8591-40f7a97e10e3" (UID: "a9740510-775e-4907-8591-40f7a97e10e3"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.494496 4799 generic.go:334] "Generic (PLEG): container finished" podID="a9740510-775e-4907-8591-40f7a97e10e3" containerID="762dacdd1ed6d04400f83778da5bc493378e06346e9026fad12bc1998c668365" exitCode=0 Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.494699 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9740510-775e-4907-8591-40f7a97e10e3","Type":"ContainerDied","Data":"762dacdd1ed6d04400f83778da5bc493378e06346e9026fad12bc1998c668365"} Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.494739 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.494810 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9740510-775e-4907-8591-40f7a97e10e3","Type":"ContainerDied","Data":"8ee07f134c04fa49d98990799f60a86f73eb656b01232b161cf1757f8f1b6f17"} Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.494840 4799 scope.go:117] "RemoveContainer" containerID="ec80651a1e3eec9a4db39805031ea01ef9f62fda7f6b36da3a77b052d1ec07da" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.524857 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9740510-775e-4907-8591-40f7a97e10e3-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.524888 4799 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9740510-775e-4907-8591-40f7a97e10e3-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.524974 4799 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a9740510-775e-4907-8591-40f7a97e10e3-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.525165 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-49tfp\" (UniqueName: \"kubernetes.io/projected/a9740510-775e-4907-8591-40f7a97e10e3-kube-api-access-49tfp\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.525176 4799 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9740510-775e-4907-8591-40f7a97e10e3-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.529948 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9740510-775e-4907-8591-40f7a97e10e3-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "a9740510-775e-4907-8591-40f7a97e10e3" (UID: "a9740510-775e-4907-8591-40f7a97e10e3"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.557031 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9740510-775e-4907-8591-40f7a97e10e3-config-data" (OuterVolumeSpecName: "config-data") pod "a9740510-775e-4907-8591-40f7a97e10e3" (UID: "a9740510-775e-4907-8591-40f7a97e10e3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.560966 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9740510-775e-4907-8591-40f7a97e10e3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a9740510-775e-4907-8591-40f7a97e10e3" (UID: "a9740510-775e-4907-8591-40f7a97e10e3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.611686 4799 scope.go:117] "RemoveContainer" containerID="0ebd7759b073d3dc35b43aae0e7f2beb369aa26b14ae4883b9f5ceb0a0451a60" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.627444 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9740510-775e-4907-8591-40f7a97e10e3-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.627479 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9740510-775e-4907-8591-40f7a97e10e3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.627493 4799 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9740510-775e-4907-8591-40f7a97e10e3-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.635644 4799 scope.go:117] "RemoveContainer" containerID="2110e05ff74ea55f482c4349e2706399889e31bbc793d354adcb30f970f95061" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.672333 4799 scope.go:117] "RemoveContainer" containerID="762dacdd1ed6d04400f83778da5bc493378e06346e9026fad12bc1998c668365" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.705846 4799 scope.go:117] "RemoveContainer" containerID="ec80651a1e3eec9a4db39805031ea01ef9f62fda7f6b36da3a77b052d1ec07da" Nov 29 05:29:39 crc kubenswrapper[4799]: E1129 05:29:39.706527 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec80651a1e3eec9a4db39805031ea01ef9f62fda7f6b36da3a77b052d1ec07da\": container with ID starting with ec80651a1e3eec9a4db39805031ea01ef9f62fda7f6b36da3a77b052d1ec07da not found: ID does not exist" containerID="ec80651a1e3eec9a4db39805031ea01ef9f62fda7f6b36da3a77b052d1ec07da" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.706584 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec80651a1e3eec9a4db39805031ea01ef9f62fda7f6b36da3a77b052d1ec07da"} err="failed to get container status \"ec80651a1e3eec9a4db39805031ea01ef9f62fda7f6b36da3a77b052d1ec07da\": rpc error: code = NotFound desc = could not find container \"ec80651a1e3eec9a4db39805031ea01ef9f62fda7f6b36da3a77b052d1ec07da\": container with ID starting with ec80651a1e3eec9a4db39805031ea01ef9f62fda7f6b36da3a77b052d1ec07da not found: ID does not exist" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.706625 4799 scope.go:117] "RemoveContainer" containerID="0ebd7759b073d3dc35b43aae0e7f2beb369aa26b14ae4883b9f5ceb0a0451a60" Nov 29 05:29:39 crc kubenswrapper[4799]: E1129 05:29:39.707223 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ebd7759b073d3dc35b43aae0e7f2beb369aa26b14ae4883b9f5ceb0a0451a60\": container with ID starting with 0ebd7759b073d3dc35b43aae0e7f2beb369aa26b14ae4883b9f5ceb0a0451a60 not found: ID does not exist" containerID="0ebd7759b073d3dc35b43aae0e7f2beb369aa26b14ae4883b9f5ceb0a0451a60" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.707261 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ebd7759b073d3dc35b43aae0e7f2beb369aa26b14ae4883b9f5ceb0a0451a60"} err="failed to get container status \"0ebd7759b073d3dc35b43aae0e7f2beb369aa26b14ae4883b9f5ceb0a0451a60\": rpc error: code = NotFound desc = could not find container \"0ebd7759b073d3dc35b43aae0e7f2beb369aa26b14ae4883b9f5ceb0a0451a60\": container with ID starting with 0ebd7759b073d3dc35b43aae0e7f2beb369aa26b14ae4883b9f5ceb0a0451a60 not found: ID does not exist" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.707295 4799 scope.go:117] "RemoveContainer" containerID="2110e05ff74ea55f482c4349e2706399889e31bbc793d354adcb30f970f95061" Nov 29 05:29:39 crc kubenswrapper[4799]: E1129 05:29:39.707526 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2110e05ff74ea55f482c4349e2706399889e31bbc793d354adcb30f970f95061\": container with ID starting with 2110e05ff74ea55f482c4349e2706399889e31bbc793d354adcb30f970f95061 not found: ID does not exist" containerID="2110e05ff74ea55f482c4349e2706399889e31bbc793d354adcb30f970f95061" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.707548 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2110e05ff74ea55f482c4349e2706399889e31bbc793d354adcb30f970f95061"} err="failed to get container status \"2110e05ff74ea55f482c4349e2706399889e31bbc793d354adcb30f970f95061\": rpc error: code = NotFound desc = could not find container \"2110e05ff74ea55f482c4349e2706399889e31bbc793d354adcb30f970f95061\": container with ID starting with 2110e05ff74ea55f482c4349e2706399889e31bbc793d354adcb30f970f95061 not found: ID does not exist" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.707561 4799 scope.go:117] "RemoveContainer" containerID="762dacdd1ed6d04400f83778da5bc493378e06346e9026fad12bc1998c668365" Nov 29 05:29:39 crc kubenswrapper[4799]: E1129 05:29:39.708004 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"762dacdd1ed6d04400f83778da5bc493378e06346e9026fad12bc1998c668365\": container with ID starting with 762dacdd1ed6d04400f83778da5bc493378e06346e9026fad12bc1998c668365 not found: ID does not exist" containerID="762dacdd1ed6d04400f83778da5bc493378e06346e9026fad12bc1998c668365" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.708056 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"762dacdd1ed6d04400f83778da5bc493378e06346e9026fad12bc1998c668365"} err="failed to get container status \"762dacdd1ed6d04400f83778da5bc493378e06346e9026fad12bc1998c668365\": rpc error: code = NotFound desc = could not find container \"762dacdd1ed6d04400f83778da5bc493378e06346e9026fad12bc1998c668365\": container with ID starting with 762dacdd1ed6d04400f83778da5bc493378e06346e9026fad12bc1998c668365 not found: ID does not exist" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.843898 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.858907 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.885289 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 29 05:29:39 crc kubenswrapper[4799]: E1129 05:29:39.886123 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee60cfe8-e5d9-4e9a-9299-38a14710fcdd" containerName="init" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.886224 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee60cfe8-e5d9-4e9a-9299-38a14710fcdd" containerName="init" Nov 29 05:29:39 crc kubenswrapper[4799]: E1129 05:29:39.886317 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9740510-775e-4907-8591-40f7a97e10e3" containerName="ceilometer-notification-agent" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.886379 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9740510-775e-4907-8591-40f7a97e10e3" containerName="ceilometer-notification-agent" Nov 29 05:29:39 crc kubenswrapper[4799]: E1129 05:29:39.886455 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9740510-775e-4907-8591-40f7a97e10e3" containerName="proxy-httpd" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.886522 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9740510-775e-4907-8591-40f7a97e10e3" containerName="proxy-httpd" Nov 29 05:29:39 crc kubenswrapper[4799]: E1129 05:29:39.886579 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9740510-775e-4907-8591-40f7a97e10e3" containerName="ceilometer-central-agent" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.886637 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9740510-775e-4907-8591-40f7a97e10e3" containerName="ceilometer-central-agent" Nov 29 05:29:39 crc kubenswrapper[4799]: E1129 05:29:39.886831 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee60cfe8-e5d9-4e9a-9299-38a14710fcdd" containerName="dnsmasq-dns" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.886906 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee60cfe8-e5d9-4e9a-9299-38a14710fcdd" containerName="dnsmasq-dns" Nov 29 05:29:39 crc kubenswrapper[4799]: E1129 05:29:39.887002 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9740510-775e-4907-8591-40f7a97e10e3" containerName="sg-core" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.887161 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9740510-775e-4907-8591-40f7a97e10e3" containerName="sg-core" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.887538 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9740510-775e-4907-8591-40f7a97e10e3" containerName="ceilometer-notification-agent" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.887635 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee60cfe8-e5d9-4e9a-9299-38a14710fcdd" containerName="dnsmasq-dns" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.887723 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9740510-775e-4907-8591-40f7a97e10e3" containerName="proxy-httpd" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.887804 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9740510-775e-4907-8591-40f7a97e10e3" containerName="ceilometer-central-agent" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.887873 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9740510-775e-4907-8591-40f7a97e10e3" containerName="sg-core" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.891145 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.896387 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.900344 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.900696 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.902349 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.934375 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c\") " pod="openstack/ceilometer-0" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.934443 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c-config-data\") pod \"ceilometer-0\" (UID: \"3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c\") " pod="openstack/ceilometer-0" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.934917 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whcmv\" (UniqueName: \"kubernetes.io/projected/3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c-kube-api-access-whcmv\") pod \"ceilometer-0\" (UID: \"3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c\") " pod="openstack/ceilometer-0" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.935053 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c-scripts\") pod \"ceilometer-0\" (UID: \"3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c\") " pod="openstack/ceilometer-0" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.935279 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c-log-httpd\") pod \"ceilometer-0\" (UID: \"3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c\") " pod="openstack/ceilometer-0" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.935353 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c\") " pod="openstack/ceilometer-0" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.935406 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c\") " pod="openstack/ceilometer-0" Nov 29 05:29:39 crc kubenswrapper[4799]: I1129 05:29:39.935453 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c-run-httpd\") pod \"ceilometer-0\" (UID: \"3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c\") " pod="openstack/ceilometer-0" Nov 29 05:29:40 crc kubenswrapper[4799]: I1129 05:29:40.037675 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whcmv\" (UniqueName: \"kubernetes.io/projected/3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c-kube-api-access-whcmv\") pod \"ceilometer-0\" (UID: \"3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c\") " pod="openstack/ceilometer-0" Nov 29 05:29:40 crc kubenswrapper[4799]: I1129 05:29:40.038072 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c-scripts\") pod \"ceilometer-0\" (UID: \"3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c\") " pod="openstack/ceilometer-0" Nov 29 05:29:40 crc kubenswrapper[4799]: I1129 05:29:40.038110 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c-log-httpd\") pod \"ceilometer-0\" (UID: \"3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c\") " pod="openstack/ceilometer-0" Nov 29 05:29:40 crc kubenswrapper[4799]: I1129 05:29:40.038132 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c\") " pod="openstack/ceilometer-0" Nov 29 05:29:40 crc kubenswrapper[4799]: I1129 05:29:40.038151 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c\") " pod="openstack/ceilometer-0" Nov 29 05:29:40 crc kubenswrapper[4799]: I1129 05:29:40.038173 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c-run-httpd\") pod \"ceilometer-0\" (UID: \"3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c\") " pod="openstack/ceilometer-0" Nov 29 05:29:40 crc kubenswrapper[4799]: I1129 05:29:40.038211 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c\") " pod="openstack/ceilometer-0" Nov 29 05:29:40 crc kubenswrapper[4799]: I1129 05:29:40.038252 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c-config-data\") pod \"ceilometer-0\" (UID: \"3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c\") " pod="openstack/ceilometer-0" Nov 29 05:29:40 crc kubenswrapper[4799]: I1129 05:29:40.038572 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c-log-httpd\") pod \"ceilometer-0\" (UID: \"3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c\") " pod="openstack/ceilometer-0" Nov 29 05:29:40 crc kubenswrapper[4799]: I1129 05:29:40.038813 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c-run-httpd\") pod \"ceilometer-0\" (UID: \"3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c\") " pod="openstack/ceilometer-0" Nov 29 05:29:40 crc kubenswrapper[4799]: I1129 05:29:40.042748 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c-scripts\") pod \"ceilometer-0\" (UID: \"3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c\") " pod="openstack/ceilometer-0" Nov 29 05:29:40 crc kubenswrapper[4799]: I1129 05:29:40.042969 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c\") " pod="openstack/ceilometer-0" Nov 29 05:29:40 crc kubenswrapper[4799]: I1129 05:29:40.043270 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c\") " pod="openstack/ceilometer-0" Nov 29 05:29:40 crc kubenswrapper[4799]: I1129 05:29:40.044443 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c\") " pod="openstack/ceilometer-0" Nov 29 05:29:40 crc kubenswrapper[4799]: I1129 05:29:40.045198 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c-config-data\") pod \"ceilometer-0\" (UID: \"3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c\") " pod="openstack/ceilometer-0" Nov 29 05:29:40 crc kubenswrapper[4799]: I1129 05:29:40.053376 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whcmv\" (UniqueName: \"kubernetes.io/projected/3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c-kube-api-access-whcmv\") pod \"ceilometer-0\" (UID: \"3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c\") " pod="openstack/ceilometer-0" Nov 29 05:29:40 crc kubenswrapper[4799]: I1129 05:29:40.213414 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 05:29:40 crc kubenswrapper[4799]: I1129 05:29:40.654348 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Nov 29 05:29:40 crc kubenswrapper[4799]: I1129 05:29:40.676752 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9740510-775e-4907-8591-40f7a97e10e3" path="/var/lib/kubelet/pods/a9740510-775e-4907-8591-40f7a97e10e3/volumes" Nov 29 05:29:40 crc kubenswrapper[4799]: W1129 05:29:40.758954 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3fadcfdb_01c3_4f90_bac4_9c30a3da6c2c.slice/crio-92924b19becbc49edb21ffc6af58044ef47582650f0097df518ff62150e46001 WatchSource:0}: Error finding container 92924b19becbc49edb21ffc6af58044ef47582650f0097df518ff62150e46001: Status 404 returned error can't find the container with id 92924b19becbc49edb21ffc6af58044ef47582650f0097df518ff62150e46001 Nov 29 05:29:40 crc kubenswrapper[4799]: I1129 05:29:40.764269 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 29 05:29:41 crc kubenswrapper[4799]: I1129 05:29:41.526744 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c","Type":"ContainerStarted","Data":"92924b19becbc49edb21ffc6af58044ef47582650f0097df518ff62150e46001"} Nov 29 05:29:42 crc kubenswrapper[4799]: I1129 05:29:42.400667 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Nov 29 05:29:42 crc kubenswrapper[4799]: I1129 05:29:42.403439 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Nov 29 05:29:42 crc kubenswrapper[4799]: I1129 05:29:42.567718 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-share-share1-0"] Nov 29 05:29:42 crc kubenswrapper[4799]: I1129 05:29:42.670271 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-share-share1-0" podUID="73c23628-dfd6-4496-9091-c1fbe3b092f6" containerName="manila-share" containerID="cri-o://bac0acaab63c16ad1ff009ee59546948c92e8b23343fb2825841b80f44e49f08" gracePeriod=30 Nov 29 05:29:42 crc kubenswrapper[4799]: I1129 05:29:42.671169 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-share-share1-0" podUID="73c23628-dfd6-4496-9091-c1fbe3b092f6" containerName="probe" containerID="cri-o://7636518fd1975296dddcdd024b7b7c7f8de8fa7adcb8789a19fe34bbcc89eb50" gracePeriod=30 Nov 29 05:29:42 crc kubenswrapper[4799]: I1129 05:29:42.756170 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c","Type":"ContainerStarted","Data":"aa22aa6fd5320a13e4af1357127c1782c92c6a1c2abdfdf6f0553e40b1a74fee"} Nov 29 05:29:42 crc kubenswrapper[4799]: I1129 05:29:42.756226 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-scheduler-0"] Nov 29 05:29:42 crc kubenswrapper[4799]: I1129 05:29:42.756475 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-scheduler-0" podUID="5d3d2b3e-1604-4813-9710-85c349aed800" containerName="manila-scheduler" containerID="cri-o://263b1dd71a80edeb22f2516a46469a2f88ea33b9ae67c572d59f9e348056324a" gracePeriod=30 Nov 29 05:29:42 crc kubenswrapper[4799]: I1129 05:29:42.756933 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-scheduler-0" podUID="5d3d2b3e-1604-4813-9710-85c349aed800" containerName="probe" containerID="cri-o://6c7a37a0fbbee07d218aa186b9370ef13bb948300d408f55ec4b3a1e234fa574" gracePeriod=30 Nov 29 05:29:43 crc kubenswrapper[4799]: I1129 05:29:43.682090 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c","Type":"ContainerStarted","Data":"9c18c03d802e338b4d08389dbba391b78205b1c63700fc7142b9d049925c6958"} Nov 29 05:29:43 crc kubenswrapper[4799]: I1129 05:29:43.684995 4799 generic.go:334] "Generic (PLEG): container finished" podID="5d3d2b3e-1604-4813-9710-85c349aed800" containerID="6c7a37a0fbbee07d218aa186b9370ef13bb948300d408f55ec4b3a1e234fa574" exitCode=0 Nov 29 05:29:43 crc kubenswrapper[4799]: I1129 05:29:43.685067 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"5d3d2b3e-1604-4813-9710-85c349aed800","Type":"ContainerDied","Data":"6c7a37a0fbbee07d218aa186b9370ef13bb948300d408f55ec4b3a1e234fa574"} Nov 29 05:29:43 crc kubenswrapper[4799]: I1129 05:29:43.693181 4799 generic.go:334] "Generic (PLEG): container finished" podID="73c23628-dfd6-4496-9091-c1fbe3b092f6" containerID="7636518fd1975296dddcdd024b7b7c7f8de8fa7adcb8789a19fe34bbcc89eb50" exitCode=0 Nov 29 05:29:43 crc kubenswrapper[4799]: I1129 05:29:43.693225 4799 generic.go:334] "Generic (PLEG): container finished" podID="73c23628-dfd6-4496-9091-c1fbe3b092f6" containerID="bac0acaab63c16ad1ff009ee59546948c92e8b23343fb2825841b80f44e49f08" exitCode=1 Nov 29 05:29:43 crc kubenswrapper[4799]: I1129 05:29:43.693254 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"73c23628-dfd6-4496-9091-c1fbe3b092f6","Type":"ContainerDied","Data":"7636518fd1975296dddcdd024b7b7c7f8de8fa7adcb8789a19fe34bbcc89eb50"} Nov 29 05:29:43 crc kubenswrapper[4799]: I1129 05:29:43.693292 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"73c23628-dfd6-4496-9091-c1fbe3b092f6","Type":"ContainerDied","Data":"bac0acaab63c16ad1ff009ee59546948c92e8b23343fb2825841b80f44e49f08"} Nov 29 05:29:43 crc kubenswrapper[4799]: I1129 05:29:43.773004 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Nov 29 05:29:43 crc kubenswrapper[4799]: I1129 05:29:43.876814 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/73c23628-dfd6-4496-9091-c1fbe3b092f6-etc-machine-id\") pod \"73c23628-dfd6-4496-9091-c1fbe3b092f6\" (UID: \"73c23628-dfd6-4496-9091-c1fbe3b092f6\") " Nov 29 05:29:43 crc kubenswrapper[4799]: I1129 05:29:43.876926 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73c23628-dfd6-4496-9091-c1fbe3b092f6-config-data\") pod \"73c23628-dfd6-4496-9091-c1fbe3b092f6\" (UID: \"73c23628-dfd6-4496-9091-c1fbe3b092f6\") " Nov 29 05:29:43 crc kubenswrapper[4799]: I1129 05:29:43.876947 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/73c23628-dfd6-4496-9091-c1fbe3b092f6-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "73c23628-dfd6-4496-9091-c1fbe3b092f6" (UID: "73c23628-dfd6-4496-9091-c1fbe3b092f6"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 05:29:43 crc kubenswrapper[4799]: I1129 05:29:43.877007 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73c23628-dfd6-4496-9091-c1fbe3b092f6-combined-ca-bundle\") pod \"73c23628-dfd6-4496-9091-c1fbe3b092f6\" (UID: \"73c23628-dfd6-4496-9091-c1fbe3b092f6\") " Nov 29 05:29:43 crc kubenswrapper[4799]: I1129 05:29:43.877071 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvnfg\" (UniqueName: \"kubernetes.io/projected/73c23628-dfd6-4496-9091-c1fbe3b092f6-kube-api-access-pvnfg\") pod \"73c23628-dfd6-4496-9091-c1fbe3b092f6\" (UID: \"73c23628-dfd6-4496-9091-c1fbe3b092f6\") " Nov 29 05:29:43 crc kubenswrapper[4799]: I1129 05:29:43.877091 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/73c23628-dfd6-4496-9091-c1fbe3b092f6-var-lib-manila\") pod \"73c23628-dfd6-4496-9091-c1fbe3b092f6\" (UID: \"73c23628-dfd6-4496-9091-c1fbe3b092f6\") " Nov 29 05:29:43 crc kubenswrapper[4799]: I1129 05:29:43.877122 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/73c23628-dfd6-4496-9091-c1fbe3b092f6-scripts\") pod \"73c23628-dfd6-4496-9091-c1fbe3b092f6\" (UID: \"73c23628-dfd6-4496-9091-c1fbe3b092f6\") " Nov 29 05:29:43 crc kubenswrapper[4799]: I1129 05:29:43.877161 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/73c23628-dfd6-4496-9091-c1fbe3b092f6-config-data-custom\") pod \"73c23628-dfd6-4496-9091-c1fbe3b092f6\" (UID: \"73c23628-dfd6-4496-9091-c1fbe3b092f6\") " Nov 29 05:29:43 crc kubenswrapper[4799]: I1129 05:29:43.877219 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/73c23628-dfd6-4496-9091-c1fbe3b092f6-ceph\") pod \"73c23628-dfd6-4496-9091-c1fbe3b092f6\" (UID: \"73c23628-dfd6-4496-9091-c1fbe3b092f6\") " Nov 29 05:29:43 crc kubenswrapper[4799]: I1129 05:29:43.877223 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/73c23628-dfd6-4496-9091-c1fbe3b092f6-var-lib-manila" (OuterVolumeSpecName: "var-lib-manila") pod "73c23628-dfd6-4496-9091-c1fbe3b092f6" (UID: "73c23628-dfd6-4496-9091-c1fbe3b092f6"). InnerVolumeSpecName "var-lib-manila". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 05:29:43 crc kubenswrapper[4799]: I1129 05:29:43.877633 4799 reconciler_common.go:293] "Volume detached for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/73c23628-dfd6-4496-9091-c1fbe3b092f6-var-lib-manila\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:43 crc kubenswrapper[4799]: I1129 05:29:43.877652 4799 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/73c23628-dfd6-4496-9091-c1fbe3b092f6-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:43 crc kubenswrapper[4799]: I1129 05:29:43.887652 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73c23628-dfd6-4496-9091-c1fbe3b092f6-kube-api-access-pvnfg" (OuterVolumeSpecName: "kube-api-access-pvnfg") pod "73c23628-dfd6-4496-9091-c1fbe3b092f6" (UID: "73c23628-dfd6-4496-9091-c1fbe3b092f6"). InnerVolumeSpecName "kube-api-access-pvnfg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:29:43 crc kubenswrapper[4799]: I1129 05:29:43.888313 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73c23628-dfd6-4496-9091-c1fbe3b092f6-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "73c23628-dfd6-4496-9091-c1fbe3b092f6" (UID: "73c23628-dfd6-4496-9091-c1fbe3b092f6"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:29:43 crc kubenswrapper[4799]: I1129 05:29:43.889028 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73c23628-dfd6-4496-9091-c1fbe3b092f6-scripts" (OuterVolumeSpecName: "scripts") pod "73c23628-dfd6-4496-9091-c1fbe3b092f6" (UID: "73c23628-dfd6-4496-9091-c1fbe3b092f6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:29:43 crc kubenswrapper[4799]: I1129 05:29:43.898433 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73c23628-dfd6-4496-9091-c1fbe3b092f6-ceph" (OuterVolumeSpecName: "ceph") pod "73c23628-dfd6-4496-9091-c1fbe3b092f6" (UID: "73c23628-dfd6-4496-9091-c1fbe3b092f6"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:29:43 crc kubenswrapper[4799]: I1129 05:29:43.943484 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73c23628-dfd6-4496-9091-c1fbe3b092f6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "73c23628-dfd6-4496-9091-c1fbe3b092f6" (UID: "73c23628-dfd6-4496-9091-c1fbe3b092f6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:29:43 crc kubenswrapper[4799]: I1129 05:29:43.980641 4799 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/73c23628-dfd6-4496-9091-c1fbe3b092f6-ceph\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:43 crc kubenswrapper[4799]: I1129 05:29:43.980767 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73c23628-dfd6-4496-9091-c1fbe3b092f6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:43 crc kubenswrapper[4799]: I1129 05:29:43.980809 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvnfg\" (UniqueName: \"kubernetes.io/projected/73c23628-dfd6-4496-9091-c1fbe3b092f6-kube-api-access-pvnfg\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:43 crc kubenswrapper[4799]: I1129 05:29:43.980823 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/73c23628-dfd6-4496-9091-c1fbe3b092f6-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:43 crc kubenswrapper[4799]: I1129 05:29:43.980837 4799 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/73c23628-dfd6-4496-9091-c1fbe3b092f6-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:44 crc kubenswrapper[4799]: I1129 05:29:44.003635 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73c23628-dfd6-4496-9091-c1fbe3b092f6-config-data" (OuterVolumeSpecName: "config-data") pod "73c23628-dfd6-4496-9091-c1fbe3b092f6" (UID: "73c23628-dfd6-4496-9091-c1fbe3b092f6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:29:44 crc kubenswrapper[4799]: I1129 05:29:44.084315 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73c23628-dfd6-4496-9091-c1fbe3b092f6-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:44 crc kubenswrapper[4799]: I1129 05:29:44.715050 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c","Type":"ContainerStarted","Data":"d81dff0800d418c653851ec18056e5340d3c49ff9cfe2e4c932389fb4ea98f7a"} Nov 29 05:29:44 crc kubenswrapper[4799]: I1129 05:29:44.717572 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"73c23628-dfd6-4496-9091-c1fbe3b092f6","Type":"ContainerDied","Data":"13d48ed228ea43f445daa4c3f940ab7e84bfeef23a264fe4e38abaa582bcf9c6"} Nov 29 05:29:44 crc kubenswrapper[4799]: I1129 05:29:44.717634 4799 scope.go:117] "RemoveContainer" containerID="7636518fd1975296dddcdd024b7b7c7f8de8fa7adcb8789a19fe34bbcc89eb50" Nov 29 05:29:44 crc kubenswrapper[4799]: I1129 05:29:44.717838 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Nov 29 05:29:44 crc kubenswrapper[4799]: I1129 05:29:44.747097 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-share-share1-0"] Nov 29 05:29:44 crc kubenswrapper[4799]: I1129 05:29:44.781690 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-share-share1-0"] Nov 29 05:29:44 crc kubenswrapper[4799]: I1129 05:29:44.798173 4799 scope.go:117] "RemoveContainer" containerID="bac0acaab63c16ad1ff009ee59546948c92e8b23343fb2825841b80f44e49f08" Nov 29 05:29:44 crc kubenswrapper[4799]: I1129 05:29:44.827402 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Nov 29 05:29:44 crc kubenswrapper[4799]: E1129 05:29:44.828039 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73c23628-dfd6-4496-9091-c1fbe3b092f6" containerName="manila-share" Nov 29 05:29:44 crc kubenswrapper[4799]: I1129 05:29:44.828056 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="73c23628-dfd6-4496-9091-c1fbe3b092f6" containerName="manila-share" Nov 29 05:29:44 crc kubenswrapper[4799]: E1129 05:29:44.828090 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73c23628-dfd6-4496-9091-c1fbe3b092f6" containerName="probe" Nov 29 05:29:44 crc kubenswrapper[4799]: I1129 05:29:44.828096 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="73c23628-dfd6-4496-9091-c1fbe3b092f6" containerName="probe" Nov 29 05:29:44 crc kubenswrapper[4799]: I1129 05:29:44.828354 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="73c23628-dfd6-4496-9091-c1fbe3b092f6" containerName="probe" Nov 29 05:29:44 crc kubenswrapper[4799]: I1129 05:29:44.828379 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="73c23628-dfd6-4496-9091-c1fbe3b092f6" containerName="manila-share" Nov 29 05:29:44 crc kubenswrapper[4799]: I1129 05:29:44.829769 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Nov 29 05:29:44 crc kubenswrapper[4799]: I1129 05:29:44.832580 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Nov 29 05:29:44 crc kubenswrapper[4799]: I1129 05:29:44.860173 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Nov 29 05:29:44 crc kubenswrapper[4799]: I1129 05:29:44.909436 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/d858c44e-0d42-4c54-b820-dc89173d67d1-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"d858c44e-0d42-4c54-b820-dc89173d67d1\") " pod="openstack/manila-share-share1-0" Nov 29 05:29:44 crc kubenswrapper[4799]: I1129 05:29:44.909922 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d858c44e-0d42-4c54-b820-dc89173d67d1-config-data\") pod \"manila-share-share1-0\" (UID: \"d858c44e-0d42-4c54-b820-dc89173d67d1\") " pod="openstack/manila-share-share1-0" Nov 29 05:29:44 crc kubenswrapper[4799]: I1129 05:29:44.910023 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d858c44e-0d42-4c54-b820-dc89173d67d1-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"d858c44e-0d42-4c54-b820-dc89173d67d1\") " pod="openstack/manila-share-share1-0" Nov 29 05:29:44 crc kubenswrapper[4799]: I1129 05:29:44.910202 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d858c44e-0d42-4c54-b820-dc89173d67d1-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"d858c44e-0d42-4c54-b820-dc89173d67d1\") " pod="openstack/manila-share-share1-0" Nov 29 05:29:44 crc kubenswrapper[4799]: I1129 05:29:44.910278 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d858c44e-0d42-4c54-b820-dc89173d67d1-ceph\") pod \"manila-share-share1-0\" (UID: \"d858c44e-0d42-4c54-b820-dc89173d67d1\") " pod="openstack/manila-share-share1-0" Nov 29 05:29:44 crc kubenswrapper[4799]: I1129 05:29:44.910443 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxssp\" (UniqueName: \"kubernetes.io/projected/d858c44e-0d42-4c54-b820-dc89173d67d1-kube-api-access-dxssp\") pod \"manila-share-share1-0\" (UID: \"d858c44e-0d42-4c54-b820-dc89173d67d1\") " pod="openstack/manila-share-share1-0" Nov 29 05:29:44 crc kubenswrapper[4799]: I1129 05:29:44.910553 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d858c44e-0d42-4c54-b820-dc89173d67d1-scripts\") pod \"manila-share-share1-0\" (UID: \"d858c44e-0d42-4c54-b820-dc89173d67d1\") " pod="openstack/manila-share-share1-0" Nov 29 05:29:44 crc kubenswrapper[4799]: I1129 05:29:44.910628 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d858c44e-0d42-4c54-b820-dc89173d67d1-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"d858c44e-0d42-4c54-b820-dc89173d67d1\") " pod="openstack/manila-share-share1-0" Nov 29 05:29:45 crc kubenswrapper[4799]: I1129 05:29:45.013208 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d858c44e-0d42-4c54-b820-dc89173d67d1-config-data\") pod \"manila-share-share1-0\" (UID: \"d858c44e-0d42-4c54-b820-dc89173d67d1\") " pod="openstack/manila-share-share1-0" Nov 29 05:29:45 crc kubenswrapper[4799]: I1129 05:29:45.013928 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d858c44e-0d42-4c54-b820-dc89173d67d1-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"d858c44e-0d42-4c54-b820-dc89173d67d1\") " pod="openstack/manila-share-share1-0" Nov 29 05:29:45 crc kubenswrapper[4799]: I1129 05:29:45.014078 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d858c44e-0d42-4c54-b820-dc89173d67d1-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"d858c44e-0d42-4c54-b820-dc89173d67d1\") " pod="openstack/manila-share-share1-0" Nov 29 05:29:45 crc kubenswrapper[4799]: I1129 05:29:45.014177 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d858c44e-0d42-4c54-b820-dc89173d67d1-ceph\") pod \"manila-share-share1-0\" (UID: \"d858c44e-0d42-4c54-b820-dc89173d67d1\") " pod="openstack/manila-share-share1-0" Nov 29 05:29:45 crc kubenswrapper[4799]: I1129 05:29:45.014233 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d858c44e-0d42-4c54-b820-dc89173d67d1-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"d858c44e-0d42-4c54-b820-dc89173d67d1\") " pod="openstack/manila-share-share1-0" Nov 29 05:29:45 crc kubenswrapper[4799]: I1129 05:29:45.014262 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxssp\" (UniqueName: \"kubernetes.io/projected/d858c44e-0d42-4c54-b820-dc89173d67d1-kube-api-access-dxssp\") pod \"manila-share-share1-0\" (UID: \"d858c44e-0d42-4c54-b820-dc89173d67d1\") " pod="openstack/manila-share-share1-0" Nov 29 05:29:45 crc kubenswrapper[4799]: I1129 05:29:45.014493 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d858c44e-0d42-4c54-b820-dc89173d67d1-scripts\") pod \"manila-share-share1-0\" (UID: \"d858c44e-0d42-4c54-b820-dc89173d67d1\") " pod="openstack/manila-share-share1-0" Nov 29 05:29:45 crc kubenswrapper[4799]: I1129 05:29:45.014539 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d858c44e-0d42-4c54-b820-dc89173d67d1-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"d858c44e-0d42-4c54-b820-dc89173d67d1\") " pod="openstack/manila-share-share1-0" Nov 29 05:29:45 crc kubenswrapper[4799]: I1129 05:29:45.014751 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/d858c44e-0d42-4c54-b820-dc89173d67d1-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"d858c44e-0d42-4c54-b820-dc89173d67d1\") " pod="openstack/manila-share-share1-0" Nov 29 05:29:45 crc kubenswrapper[4799]: I1129 05:29:45.014963 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/d858c44e-0d42-4c54-b820-dc89173d67d1-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"d858c44e-0d42-4c54-b820-dc89173d67d1\") " pod="openstack/manila-share-share1-0" Nov 29 05:29:45 crc kubenswrapper[4799]: I1129 05:29:45.018396 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d858c44e-0d42-4c54-b820-dc89173d67d1-ceph\") pod \"manila-share-share1-0\" (UID: \"d858c44e-0d42-4c54-b820-dc89173d67d1\") " pod="openstack/manila-share-share1-0" Nov 29 05:29:45 crc kubenswrapper[4799]: I1129 05:29:45.018496 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d858c44e-0d42-4c54-b820-dc89173d67d1-scripts\") pod \"manila-share-share1-0\" (UID: \"d858c44e-0d42-4c54-b820-dc89173d67d1\") " pod="openstack/manila-share-share1-0" Nov 29 05:29:45 crc kubenswrapper[4799]: I1129 05:29:45.018560 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d858c44e-0d42-4c54-b820-dc89173d67d1-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"d858c44e-0d42-4c54-b820-dc89173d67d1\") " pod="openstack/manila-share-share1-0" Nov 29 05:29:45 crc kubenswrapper[4799]: I1129 05:29:45.019959 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d858c44e-0d42-4c54-b820-dc89173d67d1-config-data\") pod \"manila-share-share1-0\" (UID: \"d858c44e-0d42-4c54-b820-dc89173d67d1\") " pod="openstack/manila-share-share1-0" Nov 29 05:29:45 crc kubenswrapper[4799]: I1129 05:29:45.032151 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d858c44e-0d42-4c54-b820-dc89173d67d1-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"d858c44e-0d42-4c54-b820-dc89173d67d1\") " pod="openstack/manila-share-share1-0" Nov 29 05:29:45 crc kubenswrapper[4799]: I1129 05:29:45.032857 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxssp\" (UniqueName: \"kubernetes.io/projected/d858c44e-0d42-4c54-b820-dc89173d67d1-kube-api-access-dxssp\") pod \"manila-share-share1-0\" (UID: \"d858c44e-0d42-4c54-b820-dc89173d67d1\") " pod="openstack/manila-share-share1-0" Nov 29 05:29:45 crc kubenswrapper[4799]: I1129 05:29:45.314236 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Nov 29 05:29:45 crc kubenswrapper[4799]: I1129 05:29:45.729162 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c","Type":"ContainerStarted","Data":"c5bc13faaec94b5d563c9d249dd3d33b80ae00d74f68ef4cf7fc49cb8029abed"} Nov 29 05:29:45 crc kubenswrapper[4799]: I1129 05:29:45.729667 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 29 05:29:45 crc kubenswrapper[4799]: I1129 05:29:45.755149 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.66288033 podStartE2EDuration="6.755121052s" podCreationTimestamp="2025-11-29 05:29:39 +0000 UTC" firstStartedPulling="2025-11-29 05:29:40.767417942 +0000 UTC m=+3056.410348382" lastFinishedPulling="2025-11-29 05:29:44.859658704 +0000 UTC m=+3060.502589104" observedRunningTime="2025-11-29 05:29:45.749263872 +0000 UTC m=+3061.392194272" watchObservedRunningTime="2025-11-29 05:29:45.755121052 +0000 UTC m=+3061.398051452" Nov 29 05:29:45 crc kubenswrapper[4799]: I1129 05:29:45.899650 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Nov 29 05:29:45 crc kubenswrapper[4799]: W1129 05:29:45.909985 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd858c44e_0d42_4c54_b820_dc89173d67d1.slice/crio-4ab5feabe9fc590e690311377b50777105eac94544ac10193df323be617d249e WatchSource:0}: Error finding container 4ab5feabe9fc590e690311377b50777105eac94544ac10193df323be617d249e: Status 404 returned error can't find the container with id 4ab5feabe9fc590e690311377b50777105eac94544ac10193df323be617d249e Nov 29 05:29:46 crc kubenswrapper[4799]: I1129 05:29:46.703519 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73c23628-dfd6-4496-9091-c1fbe3b092f6" path="/var/lib/kubelet/pods/73c23628-dfd6-4496-9091-c1fbe3b092f6/volumes" Nov 29 05:29:46 crc kubenswrapper[4799]: I1129 05:29:46.801927 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-86d9b546cb-5xxhb" podUID="f09087c5-87d5-4cd1-9a91-cbe48713d9e9" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.236:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.236:8443: connect: connection refused" Nov 29 05:29:46 crc kubenswrapper[4799]: I1129 05:29:46.810422 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"d858c44e-0d42-4c54-b820-dc89173d67d1","Type":"ContainerStarted","Data":"80147dd312a401b869539feb2c1ed8e807b19b49cebeec19d47f3d5fd6599f9d"} Nov 29 05:29:46 crc kubenswrapper[4799]: I1129 05:29:46.810469 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"d858c44e-0d42-4c54-b820-dc89173d67d1","Type":"ContainerStarted","Data":"4ab5feabe9fc590e690311377b50777105eac94544ac10193df323be617d249e"} Nov 29 05:29:47 crc kubenswrapper[4799]: I1129 05:29:47.693468 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Nov 29 05:29:47 crc kubenswrapper[4799]: I1129 05:29:47.785872 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d3d2b3e-1604-4813-9710-85c349aed800-scripts\") pod \"5d3d2b3e-1604-4813-9710-85c349aed800\" (UID: \"5d3d2b3e-1604-4813-9710-85c349aed800\") " Nov 29 05:29:47 crc kubenswrapper[4799]: I1129 05:29:47.785956 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vw6fd\" (UniqueName: \"kubernetes.io/projected/5d3d2b3e-1604-4813-9710-85c349aed800-kube-api-access-vw6fd\") pod \"5d3d2b3e-1604-4813-9710-85c349aed800\" (UID: \"5d3d2b3e-1604-4813-9710-85c349aed800\") " Nov 29 05:29:47 crc kubenswrapper[4799]: I1129 05:29:47.786049 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5d3d2b3e-1604-4813-9710-85c349aed800-etc-machine-id\") pod \"5d3d2b3e-1604-4813-9710-85c349aed800\" (UID: \"5d3d2b3e-1604-4813-9710-85c349aed800\") " Nov 29 05:29:47 crc kubenswrapper[4799]: I1129 05:29:47.786223 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5d3d2b3e-1604-4813-9710-85c349aed800-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "5d3d2b3e-1604-4813-9710-85c349aed800" (UID: "5d3d2b3e-1604-4813-9710-85c349aed800"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 05:29:47 crc kubenswrapper[4799]: I1129 05:29:47.786266 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5d3d2b3e-1604-4813-9710-85c349aed800-config-data-custom\") pod \"5d3d2b3e-1604-4813-9710-85c349aed800\" (UID: \"5d3d2b3e-1604-4813-9710-85c349aed800\") " Nov 29 05:29:47 crc kubenswrapper[4799]: I1129 05:29:47.786421 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d3d2b3e-1604-4813-9710-85c349aed800-combined-ca-bundle\") pod \"5d3d2b3e-1604-4813-9710-85c349aed800\" (UID: \"5d3d2b3e-1604-4813-9710-85c349aed800\") " Nov 29 05:29:47 crc kubenswrapper[4799]: I1129 05:29:47.787209 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d3d2b3e-1604-4813-9710-85c349aed800-config-data\") pod \"5d3d2b3e-1604-4813-9710-85c349aed800\" (UID: \"5d3d2b3e-1604-4813-9710-85c349aed800\") " Nov 29 05:29:47 crc kubenswrapper[4799]: I1129 05:29:47.788312 4799 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5d3d2b3e-1604-4813-9710-85c349aed800-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:47 crc kubenswrapper[4799]: I1129 05:29:47.794150 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d3d2b3e-1604-4813-9710-85c349aed800-kube-api-access-vw6fd" (OuterVolumeSpecName: "kube-api-access-vw6fd") pod "5d3d2b3e-1604-4813-9710-85c349aed800" (UID: "5d3d2b3e-1604-4813-9710-85c349aed800"). InnerVolumeSpecName "kube-api-access-vw6fd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:29:47 crc kubenswrapper[4799]: I1129 05:29:47.810565 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d3d2b3e-1604-4813-9710-85c349aed800-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "5d3d2b3e-1604-4813-9710-85c349aed800" (UID: "5d3d2b3e-1604-4813-9710-85c349aed800"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:29:47 crc kubenswrapper[4799]: I1129 05:29:47.816081 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d3d2b3e-1604-4813-9710-85c349aed800-scripts" (OuterVolumeSpecName: "scripts") pod "5d3d2b3e-1604-4813-9710-85c349aed800" (UID: "5d3d2b3e-1604-4813-9710-85c349aed800"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:29:47 crc kubenswrapper[4799]: I1129 05:29:47.865587 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d3d2b3e-1604-4813-9710-85c349aed800-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5d3d2b3e-1604-4813-9710-85c349aed800" (UID: "5d3d2b3e-1604-4813-9710-85c349aed800"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:29:47 crc kubenswrapper[4799]: I1129 05:29:47.894144 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d3d2b3e-1604-4813-9710-85c349aed800-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:47 crc kubenswrapper[4799]: I1129 05:29:47.894177 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vw6fd\" (UniqueName: \"kubernetes.io/projected/5d3d2b3e-1604-4813-9710-85c349aed800-kube-api-access-vw6fd\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:47 crc kubenswrapper[4799]: I1129 05:29:47.894188 4799 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5d3d2b3e-1604-4813-9710-85c349aed800-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:47 crc kubenswrapper[4799]: I1129 05:29:47.894199 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d3d2b3e-1604-4813-9710-85c349aed800-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:47 crc kubenswrapper[4799]: I1129 05:29:47.923313 4799 generic.go:334] "Generic (PLEG): container finished" podID="5d3d2b3e-1604-4813-9710-85c349aed800" containerID="263b1dd71a80edeb22f2516a46469a2f88ea33b9ae67c572d59f9e348056324a" exitCode=0 Nov 29 05:29:47 crc kubenswrapper[4799]: I1129 05:29:47.923475 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"5d3d2b3e-1604-4813-9710-85c349aed800","Type":"ContainerDied","Data":"263b1dd71a80edeb22f2516a46469a2f88ea33b9ae67c572d59f9e348056324a"} Nov 29 05:29:47 crc kubenswrapper[4799]: I1129 05:29:47.923518 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"5d3d2b3e-1604-4813-9710-85c349aed800","Type":"ContainerDied","Data":"ca955360d61a44a4c5676feaad11990950059dab400f30479e9e54b38507551f"} Nov 29 05:29:47 crc kubenswrapper[4799]: I1129 05:29:47.923548 4799 scope.go:117] "RemoveContainer" containerID="6c7a37a0fbbee07d218aa186b9370ef13bb948300d408f55ec4b3a1e234fa574" Nov 29 05:29:47 crc kubenswrapper[4799]: I1129 05:29:47.923817 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Nov 29 05:29:47 crc kubenswrapper[4799]: I1129 05:29:47.928278 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"d858c44e-0d42-4c54-b820-dc89173d67d1","Type":"ContainerStarted","Data":"a9271afb9cb7a6a754d1903200c0251f04bd4838c8fb64be436d5c79dfe09279"} Nov 29 05:29:47 crc kubenswrapper[4799]: I1129 05:29:47.962243 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=3.962222169 podStartE2EDuration="3.962222169s" podCreationTimestamp="2025-11-29 05:29:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 05:29:47.955239432 +0000 UTC m=+3063.598169832" watchObservedRunningTime="2025-11-29 05:29:47.962222169 +0000 UTC m=+3063.605152569" Nov 29 05:29:47 crc kubenswrapper[4799]: I1129 05:29:47.966958 4799 scope.go:117] "RemoveContainer" containerID="263b1dd71a80edeb22f2516a46469a2f88ea33b9ae67c572d59f9e348056324a" Nov 29 05:29:47 crc kubenswrapper[4799]: I1129 05:29:47.972965 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d3d2b3e-1604-4813-9710-85c349aed800-config-data" (OuterVolumeSpecName: "config-data") pod "5d3d2b3e-1604-4813-9710-85c349aed800" (UID: "5d3d2b3e-1604-4813-9710-85c349aed800"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:29:47 crc kubenswrapper[4799]: I1129 05:29:47.987947 4799 scope.go:117] "RemoveContainer" containerID="6c7a37a0fbbee07d218aa186b9370ef13bb948300d408f55ec4b3a1e234fa574" Nov 29 05:29:47 crc kubenswrapper[4799]: E1129 05:29:47.988945 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c7a37a0fbbee07d218aa186b9370ef13bb948300d408f55ec4b3a1e234fa574\": container with ID starting with 6c7a37a0fbbee07d218aa186b9370ef13bb948300d408f55ec4b3a1e234fa574 not found: ID does not exist" containerID="6c7a37a0fbbee07d218aa186b9370ef13bb948300d408f55ec4b3a1e234fa574" Nov 29 05:29:47 crc kubenswrapper[4799]: I1129 05:29:47.988997 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c7a37a0fbbee07d218aa186b9370ef13bb948300d408f55ec4b3a1e234fa574"} err="failed to get container status \"6c7a37a0fbbee07d218aa186b9370ef13bb948300d408f55ec4b3a1e234fa574\": rpc error: code = NotFound desc = could not find container \"6c7a37a0fbbee07d218aa186b9370ef13bb948300d408f55ec4b3a1e234fa574\": container with ID starting with 6c7a37a0fbbee07d218aa186b9370ef13bb948300d408f55ec4b3a1e234fa574 not found: ID does not exist" Nov 29 05:29:47 crc kubenswrapper[4799]: I1129 05:29:47.989033 4799 scope.go:117] "RemoveContainer" containerID="263b1dd71a80edeb22f2516a46469a2f88ea33b9ae67c572d59f9e348056324a" Nov 29 05:29:47 crc kubenswrapper[4799]: E1129 05:29:47.989431 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"263b1dd71a80edeb22f2516a46469a2f88ea33b9ae67c572d59f9e348056324a\": container with ID starting with 263b1dd71a80edeb22f2516a46469a2f88ea33b9ae67c572d59f9e348056324a not found: ID does not exist" containerID="263b1dd71a80edeb22f2516a46469a2f88ea33b9ae67c572d59f9e348056324a" Nov 29 05:29:47 crc kubenswrapper[4799]: I1129 05:29:47.989458 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"263b1dd71a80edeb22f2516a46469a2f88ea33b9ae67c572d59f9e348056324a"} err="failed to get container status \"263b1dd71a80edeb22f2516a46469a2f88ea33b9ae67c572d59f9e348056324a\": rpc error: code = NotFound desc = could not find container \"263b1dd71a80edeb22f2516a46469a2f88ea33b9ae67c572d59f9e348056324a\": container with ID starting with 263b1dd71a80edeb22f2516a46469a2f88ea33b9ae67c572d59f9e348056324a not found: ID does not exist" Nov 29 05:29:47 crc kubenswrapper[4799]: I1129 05:29:47.996705 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d3d2b3e-1604-4813-9710-85c349aed800-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 05:29:48 crc kubenswrapper[4799]: I1129 05:29:48.269343 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-scheduler-0"] Nov 29 05:29:48 crc kubenswrapper[4799]: I1129 05:29:48.288401 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-scheduler-0"] Nov 29 05:29:48 crc kubenswrapper[4799]: I1129 05:29:48.302002 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Nov 29 05:29:48 crc kubenswrapper[4799]: E1129 05:29:48.302830 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d3d2b3e-1604-4813-9710-85c349aed800" containerName="manila-scheduler" Nov 29 05:29:48 crc kubenswrapper[4799]: I1129 05:29:48.302932 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d3d2b3e-1604-4813-9710-85c349aed800" containerName="manila-scheduler" Nov 29 05:29:48 crc kubenswrapper[4799]: E1129 05:29:48.303069 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d3d2b3e-1604-4813-9710-85c349aed800" containerName="probe" Nov 29 05:29:48 crc kubenswrapper[4799]: I1129 05:29:48.303168 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d3d2b3e-1604-4813-9710-85c349aed800" containerName="probe" Nov 29 05:29:48 crc kubenswrapper[4799]: I1129 05:29:48.303514 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d3d2b3e-1604-4813-9710-85c349aed800" containerName="probe" Nov 29 05:29:48 crc kubenswrapper[4799]: I1129 05:29:48.303650 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d3d2b3e-1604-4813-9710-85c349aed800" containerName="manila-scheduler" Nov 29 05:29:48 crc kubenswrapper[4799]: I1129 05:29:48.306719 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Nov 29 05:29:48 crc kubenswrapper[4799]: I1129 05:29:48.311371 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Nov 29 05:29:48 crc kubenswrapper[4799]: I1129 05:29:48.319415 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Nov 29 05:29:48 crc kubenswrapper[4799]: I1129 05:29:48.407825 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef775c01-8e04-4efe-8192-889ddf0ccbd3-scripts\") pod \"manila-scheduler-0\" (UID: \"ef775c01-8e04-4efe-8192-889ddf0ccbd3\") " pod="openstack/manila-scheduler-0" Nov 29 05:29:48 crc kubenswrapper[4799]: I1129 05:29:48.407911 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef775c01-8e04-4efe-8192-889ddf0ccbd3-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"ef775c01-8e04-4efe-8192-889ddf0ccbd3\") " pod="openstack/manila-scheduler-0" Nov 29 05:29:48 crc kubenswrapper[4799]: I1129 05:29:48.408016 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef775c01-8e04-4efe-8192-889ddf0ccbd3-config-data\") pod \"manila-scheduler-0\" (UID: \"ef775c01-8e04-4efe-8192-889ddf0ccbd3\") " pod="openstack/manila-scheduler-0" Nov 29 05:29:48 crc kubenswrapper[4799]: I1129 05:29:48.408049 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gvhw\" (UniqueName: \"kubernetes.io/projected/ef775c01-8e04-4efe-8192-889ddf0ccbd3-kube-api-access-2gvhw\") pod \"manila-scheduler-0\" (UID: \"ef775c01-8e04-4efe-8192-889ddf0ccbd3\") " pod="openstack/manila-scheduler-0" Nov 29 05:29:48 crc kubenswrapper[4799]: I1129 05:29:48.408115 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ef775c01-8e04-4efe-8192-889ddf0ccbd3-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"ef775c01-8e04-4efe-8192-889ddf0ccbd3\") " pod="openstack/manila-scheduler-0" Nov 29 05:29:48 crc kubenswrapper[4799]: I1129 05:29:48.408149 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ef775c01-8e04-4efe-8192-889ddf0ccbd3-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"ef775c01-8e04-4efe-8192-889ddf0ccbd3\") " pod="openstack/manila-scheduler-0" Nov 29 05:29:48 crc kubenswrapper[4799]: I1129 05:29:48.509951 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef775c01-8e04-4efe-8192-889ddf0ccbd3-scripts\") pod \"manila-scheduler-0\" (UID: \"ef775c01-8e04-4efe-8192-889ddf0ccbd3\") " pod="openstack/manila-scheduler-0" Nov 29 05:29:48 crc kubenswrapper[4799]: I1129 05:29:48.510112 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef775c01-8e04-4efe-8192-889ddf0ccbd3-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"ef775c01-8e04-4efe-8192-889ddf0ccbd3\") " pod="openstack/manila-scheduler-0" Nov 29 05:29:48 crc kubenswrapper[4799]: I1129 05:29:48.510245 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef775c01-8e04-4efe-8192-889ddf0ccbd3-config-data\") pod \"manila-scheduler-0\" (UID: \"ef775c01-8e04-4efe-8192-889ddf0ccbd3\") " pod="openstack/manila-scheduler-0" Nov 29 05:29:48 crc kubenswrapper[4799]: I1129 05:29:48.510290 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gvhw\" (UniqueName: \"kubernetes.io/projected/ef775c01-8e04-4efe-8192-889ddf0ccbd3-kube-api-access-2gvhw\") pod \"manila-scheduler-0\" (UID: \"ef775c01-8e04-4efe-8192-889ddf0ccbd3\") " pod="openstack/manila-scheduler-0" Nov 29 05:29:48 crc kubenswrapper[4799]: I1129 05:29:48.510365 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ef775c01-8e04-4efe-8192-889ddf0ccbd3-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"ef775c01-8e04-4efe-8192-889ddf0ccbd3\") " pod="openstack/manila-scheduler-0" Nov 29 05:29:48 crc kubenswrapper[4799]: I1129 05:29:48.510411 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ef775c01-8e04-4efe-8192-889ddf0ccbd3-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"ef775c01-8e04-4efe-8192-889ddf0ccbd3\") " pod="openstack/manila-scheduler-0" Nov 29 05:29:48 crc kubenswrapper[4799]: I1129 05:29:48.510633 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ef775c01-8e04-4efe-8192-889ddf0ccbd3-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"ef775c01-8e04-4efe-8192-889ddf0ccbd3\") " pod="openstack/manila-scheduler-0" Nov 29 05:29:48 crc kubenswrapper[4799]: I1129 05:29:48.514137 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef775c01-8e04-4efe-8192-889ddf0ccbd3-scripts\") pod \"manila-scheduler-0\" (UID: \"ef775c01-8e04-4efe-8192-889ddf0ccbd3\") " pod="openstack/manila-scheduler-0" Nov 29 05:29:48 crc kubenswrapper[4799]: I1129 05:29:48.515977 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ef775c01-8e04-4efe-8192-889ddf0ccbd3-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"ef775c01-8e04-4efe-8192-889ddf0ccbd3\") " pod="openstack/manila-scheduler-0" Nov 29 05:29:48 crc kubenswrapper[4799]: I1129 05:29:48.517091 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef775c01-8e04-4efe-8192-889ddf0ccbd3-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"ef775c01-8e04-4efe-8192-889ddf0ccbd3\") " pod="openstack/manila-scheduler-0" Nov 29 05:29:48 crc kubenswrapper[4799]: I1129 05:29:48.520937 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef775c01-8e04-4efe-8192-889ddf0ccbd3-config-data\") pod \"manila-scheduler-0\" (UID: \"ef775c01-8e04-4efe-8192-889ddf0ccbd3\") " pod="openstack/manila-scheduler-0" Nov 29 05:29:48 crc kubenswrapper[4799]: I1129 05:29:48.530083 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gvhw\" (UniqueName: \"kubernetes.io/projected/ef775c01-8e04-4efe-8192-889ddf0ccbd3-kube-api-access-2gvhw\") pod \"manila-scheduler-0\" (UID: \"ef775c01-8e04-4efe-8192-889ddf0ccbd3\") " pod="openstack/manila-scheduler-0" Nov 29 05:29:48 crc kubenswrapper[4799]: I1129 05:29:48.629859 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Nov 29 05:29:48 crc kubenswrapper[4799]: I1129 05:29:48.661062 4799 scope.go:117] "RemoveContainer" containerID="e6d4b3d1d5b5e1eeb13508d67de3d3c3aac52e76337b55e53793185c0d001d82" Nov 29 05:29:48 crc kubenswrapper[4799]: E1129 05:29:48.661358 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:29:48 crc kubenswrapper[4799]: I1129 05:29:48.674236 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d3d2b3e-1604-4813-9710-85c349aed800" path="/var/lib/kubelet/pods/5d3d2b3e-1604-4813-9710-85c349aed800/volumes" Nov 29 05:29:49 crc kubenswrapper[4799]: I1129 05:29:49.136092 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/manila-api-0" Nov 29 05:29:49 crc kubenswrapper[4799]: I1129 05:29:49.142481 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Nov 29 05:29:49 crc kubenswrapper[4799]: I1129 05:29:49.959381 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"ef775c01-8e04-4efe-8192-889ddf0ccbd3","Type":"ContainerStarted","Data":"b29d0a3c249a77ca31ec0ad402022af9f571b06a086803043ddbb739fb0374cd"} Nov 29 05:29:49 crc kubenswrapper[4799]: I1129 05:29:49.960758 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"ef775c01-8e04-4efe-8192-889ddf0ccbd3","Type":"ContainerStarted","Data":"459f27152046ab5e02daf530de30d7a5d02333a2a1d3bf647c7c9ce3a8243044"} Nov 29 05:29:50 crc kubenswrapper[4799]: I1129 05:29:50.971328 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"ef775c01-8e04-4efe-8192-889ddf0ccbd3","Type":"ContainerStarted","Data":"7e3337661554301791a35fda8e5868c53f997a9bdccfea58a1b06244beac6be0"} Nov 29 05:29:51 crc kubenswrapper[4799]: I1129 05:29:51.000976 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=3.000947123 podStartE2EDuration="3.000947123s" podCreationTimestamp="2025-11-29 05:29:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 05:29:50.993202647 +0000 UTC m=+3066.636133047" watchObservedRunningTime="2025-11-29 05:29:51.000947123 +0000 UTC m=+3066.643877533" Nov 29 05:29:55 crc kubenswrapper[4799]: I1129 05:29:55.315086 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Nov 29 05:29:56 crc kubenswrapper[4799]: I1129 05:29:56.798582 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-86d9b546cb-5xxhb" podUID="f09087c5-87d5-4cd1-9a91-cbe48713d9e9" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.236:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.236:8443: connect: connection refused" Nov 29 05:29:58 crc kubenswrapper[4799]: I1129 05:29:58.630479 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Nov 29 05:30:00 crc kubenswrapper[4799]: I1129 05:30:00.174027 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406570-r4g4v"] Nov 29 05:30:00 crc kubenswrapper[4799]: I1129 05:30:00.177223 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406570-r4g4v" Nov 29 05:30:00 crc kubenswrapper[4799]: I1129 05:30:00.182234 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 29 05:30:00 crc kubenswrapper[4799]: I1129 05:30:00.182275 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 29 05:30:00 crc kubenswrapper[4799]: I1129 05:30:00.187694 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406570-r4g4v"] Nov 29 05:30:00 crc kubenswrapper[4799]: I1129 05:30:00.305701 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f211bc8c-978a-476b-b7e1-76bb4c6ef18d-secret-volume\") pod \"collect-profiles-29406570-r4g4v\" (UID: \"f211bc8c-978a-476b-b7e1-76bb4c6ef18d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406570-r4g4v" Nov 29 05:30:00 crc kubenswrapper[4799]: I1129 05:30:00.306305 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f211bc8c-978a-476b-b7e1-76bb4c6ef18d-config-volume\") pod \"collect-profiles-29406570-r4g4v\" (UID: \"f211bc8c-978a-476b-b7e1-76bb4c6ef18d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406570-r4g4v" Nov 29 05:30:00 crc kubenswrapper[4799]: I1129 05:30:00.306403 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktrcg\" (UniqueName: \"kubernetes.io/projected/f211bc8c-978a-476b-b7e1-76bb4c6ef18d-kube-api-access-ktrcg\") pod \"collect-profiles-29406570-r4g4v\" (UID: \"f211bc8c-978a-476b-b7e1-76bb4c6ef18d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406570-r4g4v" Nov 29 05:30:00 crc kubenswrapper[4799]: I1129 05:30:00.410065 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktrcg\" (UniqueName: \"kubernetes.io/projected/f211bc8c-978a-476b-b7e1-76bb4c6ef18d-kube-api-access-ktrcg\") pod \"collect-profiles-29406570-r4g4v\" (UID: \"f211bc8c-978a-476b-b7e1-76bb4c6ef18d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406570-r4g4v" Nov 29 05:30:00 crc kubenswrapper[4799]: I1129 05:30:00.410387 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f211bc8c-978a-476b-b7e1-76bb4c6ef18d-secret-volume\") pod \"collect-profiles-29406570-r4g4v\" (UID: \"f211bc8c-978a-476b-b7e1-76bb4c6ef18d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406570-r4g4v" Nov 29 05:30:00 crc kubenswrapper[4799]: I1129 05:30:00.410590 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f211bc8c-978a-476b-b7e1-76bb4c6ef18d-config-volume\") pod \"collect-profiles-29406570-r4g4v\" (UID: \"f211bc8c-978a-476b-b7e1-76bb4c6ef18d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406570-r4g4v" Nov 29 05:30:00 crc kubenswrapper[4799]: I1129 05:30:00.412898 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f211bc8c-978a-476b-b7e1-76bb4c6ef18d-config-volume\") pod \"collect-profiles-29406570-r4g4v\" (UID: \"f211bc8c-978a-476b-b7e1-76bb4c6ef18d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406570-r4g4v" Nov 29 05:30:00 crc kubenswrapper[4799]: I1129 05:30:00.436619 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f211bc8c-978a-476b-b7e1-76bb4c6ef18d-secret-volume\") pod \"collect-profiles-29406570-r4g4v\" (UID: \"f211bc8c-978a-476b-b7e1-76bb4c6ef18d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406570-r4g4v" Nov 29 05:30:00 crc kubenswrapper[4799]: I1129 05:30:00.439299 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktrcg\" (UniqueName: \"kubernetes.io/projected/f211bc8c-978a-476b-b7e1-76bb4c6ef18d-kube-api-access-ktrcg\") pod \"collect-profiles-29406570-r4g4v\" (UID: \"f211bc8c-978a-476b-b7e1-76bb4c6ef18d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406570-r4g4v" Nov 29 05:30:00 crc kubenswrapper[4799]: I1129 05:30:00.507997 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406570-r4g4v" Nov 29 05:30:01 crc kubenswrapper[4799]: I1129 05:30:01.083030 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406570-r4g4v"] Nov 29 05:30:01 crc kubenswrapper[4799]: I1129 05:30:01.133540 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29406570-r4g4v" event={"ID":"f211bc8c-978a-476b-b7e1-76bb4c6ef18d","Type":"ContainerStarted","Data":"06b837a5e377bd20f5c2b72a1996b39470d3da1ca29862a77e2e4afb319632dd"} Nov 29 05:30:01 crc kubenswrapper[4799]: I1129 05:30:01.659747 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-86d9b546cb-5xxhb" Nov 29 05:30:01 crc kubenswrapper[4799]: I1129 05:30:01.750998 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f09087c5-87d5-4cd1-9a91-cbe48713d9e9-logs\") pod \"f09087c5-87d5-4cd1-9a91-cbe48713d9e9\" (UID: \"f09087c5-87d5-4cd1-9a91-cbe48713d9e9\") " Nov 29 05:30:01 crc kubenswrapper[4799]: I1129 05:30:01.751568 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f09087c5-87d5-4cd1-9a91-cbe48713d9e9-logs" (OuterVolumeSpecName: "logs") pod "f09087c5-87d5-4cd1-9a91-cbe48713d9e9" (UID: "f09087c5-87d5-4cd1-9a91-cbe48713d9e9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:30:01 crc kubenswrapper[4799]: I1129 05:30:01.751668 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f09087c5-87d5-4cd1-9a91-cbe48713d9e9-scripts\") pod \"f09087c5-87d5-4cd1-9a91-cbe48713d9e9\" (UID: \"f09087c5-87d5-4cd1-9a91-cbe48713d9e9\") " Nov 29 05:30:01 crc kubenswrapper[4799]: I1129 05:30:01.751734 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f09087c5-87d5-4cd1-9a91-cbe48713d9e9-horizon-secret-key\") pod \"f09087c5-87d5-4cd1-9a91-cbe48713d9e9\" (UID: \"f09087c5-87d5-4cd1-9a91-cbe48713d9e9\") " Nov 29 05:30:01 crc kubenswrapper[4799]: I1129 05:30:01.752070 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/f09087c5-87d5-4cd1-9a91-cbe48713d9e9-horizon-tls-certs\") pod \"f09087c5-87d5-4cd1-9a91-cbe48713d9e9\" (UID: \"f09087c5-87d5-4cd1-9a91-cbe48713d9e9\") " Nov 29 05:30:01 crc kubenswrapper[4799]: I1129 05:30:01.752155 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4b5g\" (UniqueName: \"kubernetes.io/projected/f09087c5-87d5-4cd1-9a91-cbe48713d9e9-kube-api-access-p4b5g\") pod \"f09087c5-87d5-4cd1-9a91-cbe48713d9e9\" (UID: \"f09087c5-87d5-4cd1-9a91-cbe48713d9e9\") " Nov 29 05:30:01 crc kubenswrapper[4799]: I1129 05:30:01.752260 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f09087c5-87d5-4cd1-9a91-cbe48713d9e9-config-data\") pod \"f09087c5-87d5-4cd1-9a91-cbe48713d9e9\" (UID: \"f09087c5-87d5-4cd1-9a91-cbe48713d9e9\") " Nov 29 05:30:01 crc kubenswrapper[4799]: I1129 05:30:01.752298 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f09087c5-87d5-4cd1-9a91-cbe48713d9e9-combined-ca-bundle\") pod \"f09087c5-87d5-4cd1-9a91-cbe48713d9e9\" (UID: \"f09087c5-87d5-4cd1-9a91-cbe48713d9e9\") " Nov 29 05:30:01 crc kubenswrapper[4799]: I1129 05:30:01.753314 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f09087c5-87d5-4cd1-9a91-cbe48713d9e9-logs\") on node \"crc\" DevicePath \"\"" Nov 29 05:30:01 crc kubenswrapper[4799]: I1129 05:30:01.759167 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f09087c5-87d5-4cd1-9a91-cbe48713d9e9-kube-api-access-p4b5g" (OuterVolumeSpecName: "kube-api-access-p4b5g") pod "f09087c5-87d5-4cd1-9a91-cbe48713d9e9" (UID: "f09087c5-87d5-4cd1-9a91-cbe48713d9e9"). InnerVolumeSpecName "kube-api-access-p4b5g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:30:01 crc kubenswrapper[4799]: I1129 05:30:01.759990 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f09087c5-87d5-4cd1-9a91-cbe48713d9e9-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "f09087c5-87d5-4cd1-9a91-cbe48713d9e9" (UID: "f09087c5-87d5-4cd1-9a91-cbe48713d9e9"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:30:01 crc kubenswrapper[4799]: I1129 05:30:01.784640 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f09087c5-87d5-4cd1-9a91-cbe48713d9e9-config-data" (OuterVolumeSpecName: "config-data") pod "f09087c5-87d5-4cd1-9a91-cbe48713d9e9" (UID: "f09087c5-87d5-4cd1-9a91-cbe48713d9e9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 05:30:01 crc kubenswrapper[4799]: I1129 05:30:01.790393 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f09087c5-87d5-4cd1-9a91-cbe48713d9e9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f09087c5-87d5-4cd1-9a91-cbe48713d9e9" (UID: "f09087c5-87d5-4cd1-9a91-cbe48713d9e9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:30:01 crc kubenswrapper[4799]: I1129 05:30:01.792220 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f09087c5-87d5-4cd1-9a91-cbe48713d9e9-scripts" (OuterVolumeSpecName: "scripts") pod "f09087c5-87d5-4cd1-9a91-cbe48713d9e9" (UID: "f09087c5-87d5-4cd1-9a91-cbe48713d9e9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 05:30:01 crc kubenswrapper[4799]: I1129 05:30:01.808473 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f09087c5-87d5-4cd1-9a91-cbe48713d9e9-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "f09087c5-87d5-4cd1-9a91-cbe48713d9e9" (UID: "f09087c5-87d5-4cd1-9a91-cbe48713d9e9"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:30:01 crc kubenswrapper[4799]: I1129 05:30:01.855893 4799 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f09087c5-87d5-4cd1-9a91-cbe48713d9e9-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 29 05:30:01 crc kubenswrapper[4799]: I1129 05:30:01.855945 4799 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/f09087c5-87d5-4cd1-9a91-cbe48713d9e9-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 29 05:30:01 crc kubenswrapper[4799]: I1129 05:30:01.855964 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4b5g\" (UniqueName: \"kubernetes.io/projected/f09087c5-87d5-4cd1-9a91-cbe48713d9e9-kube-api-access-p4b5g\") on node \"crc\" DevicePath \"\"" Nov 29 05:30:01 crc kubenswrapper[4799]: I1129 05:30:01.855990 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f09087c5-87d5-4cd1-9a91-cbe48713d9e9-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 05:30:01 crc kubenswrapper[4799]: I1129 05:30:01.856006 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f09087c5-87d5-4cd1-9a91-cbe48713d9e9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 05:30:01 crc kubenswrapper[4799]: I1129 05:30:01.856021 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f09087c5-87d5-4cd1-9a91-cbe48713d9e9-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 05:30:02 crc kubenswrapper[4799]: I1129 05:30:02.148749 4799 generic.go:334] "Generic (PLEG): container finished" podID="f09087c5-87d5-4cd1-9a91-cbe48713d9e9" containerID="8f1f666f1cb9af792d04255e0b4e5900be38a9d0b4060fd4c97f314b6c12a164" exitCode=137 Nov 29 05:30:02 crc kubenswrapper[4799]: I1129 05:30:02.148858 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-86d9b546cb-5xxhb" event={"ID":"f09087c5-87d5-4cd1-9a91-cbe48713d9e9","Type":"ContainerDied","Data":"8f1f666f1cb9af792d04255e0b4e5900be38a9d0b4060fd4c97f314b6c12a164"} Nov 29 05:30:02 crc kubenswrapper[4799]: I1129 05:30:02.148895 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-86d9b546cb-5xxhb" event={"ID":"f09087c5-87d5-4cd1-9a91-cbe48713d9e9","Type":"ContainerDied","Data":"c930f52404aeb6dd443f7a9afd2c5b13d6e72dac127362739964a73f6fff9441"} Nov 29 05:30:02 crc kubenswrapper[4799]: I1129 05:30:02.148926 4799 scope.go:117] "RemoveContainer" containerID="acf62f8dda4e17622de5c459a6b4fe1c70a55e3481d77ef29c8a3bc868f9df26" Nov 29 05:30:02 crc kubenswrapper[4799]: I1129 05:30:02.149073 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-86d9b546cb-5xxhb" Nov 29 05:30:02 crc kubenswrapper[4799]: I1129 05:30:02.157044 4799 generic.go:334] "Generic (PLEG): container finished" podID="f211bc8c-978a-476b-b7e1-76bb4c6ef18d" containerID="91c8fe691b686c562e413547338becb63396dd4142a3b71d12d7da2c68440fec" exitCode=0 Nov 29 05:30:02 crc kubenswrapper[4799]: I1129 05:30:02.157084 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29406570-r4g4v" event={"ID":"f211bc8c-978a-476b-b7e1-76bb4c6ef18d","Type":"ContainerDied","Data":"91c8fe691b686c562e413547338becb63396dd4142a3b71d12d7da2c68440fec"} Nov 29 05:30:02 crc kubenswrapper[4799]: I1129 05:30:02.225994 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-86d9b546cb-5xxhb"] Nov 29 05:30:02 crc kubenswrapper[4799]: I1129 05:30:02.240817 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-86d9b546cb-5xxhb"] Nov 29 05:30:02 crc kubenswrapper[4799]: I1129 05:30:02.366167 4799 scope.go:117] "RemoveContainer" containerID="8f1f666f1cb9af792d04255e0b4e5900be38a9d0b4060fd4c97f314b6c12a164" Nov 29 05:30:02 crc kubenswrapper[4799]: I1129 05:30:02.391622 4799 scope.go:117] "RemoveContainer" containerID="acf62f8dda4e17622de5c459a6b4fe1c70a55e3481d77ef29c8a3bc868f9df26" Nov 29 05:30:02 crc kubenswrapper[4799]: E1129 05:30:02.392341 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"acf62f8dda4e17622de5c459a6b4fe1c70a55e3481d77ef29c8a3bc868f9df26\": container with ID starting with acf62f8dda4e17622de5c459a6b4fe1c70a55e3481d77ef29c8a3bc868f9df26 not found: ID does not exist" containerID="acf62f8dda4e17622de5c459a6b4fe1c70a55e3481d77ef29c8a3bc868f9df26" Nov 29 05:30:02 crc kubenswrapper[4799]: I1129 05:30:02.392385 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"acf62f8dda4e17622de5c459a6b4fe1c70a55e3481d77ef29c8a3bc868f9df26"} err="failed to get container status \"acf62f8dda4e17622de5c459a6b4fe1c70a55e3481d77ef29c8a3bc868f9df26\": rpc error: code = NotFound desc = could not find container \"acf62f8dda4e17622de5c459a6b4fe1c70a55e3481d77ef29c8a3bc868f9df26\": container with ID starting with acf62f8dda4e17622de5c459a6b4fe1c70a55e3481d77ef29c8a3bc868f9df26 not found: ID does not exist" Nov 29 05:30:02 crc kubenswrapper[4799]: I1129 05:30:02.392415 4799 scope.go:117] "RemoveContainer" containerID="8f1f666f1cb9af792d04255e0b4e5900be38a9d0b4060fd4c97f314b6c12a164" Nov 29 05:30:02 crc kubenswrapper[4799]: E1129 05:30:02.392770 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f1f666f1cb9af792d04255e0b4e5900be38a9d0b4060fd4c97f314b6c12a164\": container with ID starting with 8f1f666f1cb9af792d04255e0b4e5900be38a9d0b4060fd4c97f314b6c12a164 not found: ID does not exist" containerID="8f1f666f1cb9af792d04255e0b4e5900be38a9d0b4060fd4c97f314b6c12a164" Nov 29 05:30:02 crc kubenswrapper[4799]: I1129 05:30:02.392812 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f1f666f1cb9af792d04255e0b4e5900be38a9d0b4060fd4c97f314b6c12a164"} err="failed to get container status \"8f1f666f1cb9af792d04255e0b4e5900be38a9d0b4060fd4c97f314b6c12a164\": rpc error: code = NotFound desc = could not find container \"8f1f666f1cb9af792d04255e0b4e5900be38a9d0b4060fd4c97f314b6c12a164\": container with ID starting with 8f1f666f1cb9af792d04255e0b4e5900be38a9d0b4060fd4c97f314b6c12a164 not found: ID does not exist" Nov 29 05:30:02 crc kubenswrapper[4799]: I1129 05:30:02.659096 4799 scope.go:117] "RemoveContainer" containerID="e6d4b3d1d5b5e1eeb13508d67de3d3c3aac52e76337b55e53793185c0d001d82" Nov 29 05:30:02 crc kubenswrapper[4799]: E1129 05:30:02.659501 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:30:02 crc kubenswrapper[4799]: I1129 05:30:02.672981 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f09087c5-87d5-4cd1-9a91-cbe48713d9e9" path="/var/lib/kubelet/pods/f09087c5-87d5-4cd1-9a91-cbe48713d9e9/volumes" Nov 29 05:30:03 crc kubenswrapper[4799]: I1129 05:30:03.529584 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406570-r4g4v" Nov 29 05:30:03 crc kubenswrapper[4799]: I1129 05:30:03.590352 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f211bc8c-978a-476b-b7e1-76bb4c6ef18d-config-volume\") pod \"f211bc8c-978a-476b-b7e1-76bb4c6ef18d\" (UID: \"f211bc8c-978a-476b-b7e1-76bb4c6ef18d\") " Nov 29 05:30:03 crc kubenswrapper[4799]: I1129 05:30:03.590454 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f211bc8c-978a-476b-b7e1-76bb4c6ef18d-secret-volume\") pod \"f211bc8c-978a-476b-b7e1-76bb4c6ef18d\" (UID: \"f211bc8c-978a-476b-b7e1-76bb4c6ef18d\") " Nov 29 05:30:03 crc kubenswrapper[4799]: I1129 05:30:03.590763 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ktrcg\" (UniqueName: \"kubernetes.io/projected/f211bc8c-978a-476b-b7e1-76bb4c6ef18d-kube-api-access-ktrcg\") pod \"f211bc8c-978a-476b-b7e1-76bb4c6ef18d\" (UID: \"f211bc8c-978a-476b-b7e1-76bb4c6ef18d\") " Nov 29 05:30:03 crc kubenswrapper[4799]: I1129 05:30:03.591243 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f211bc8c-978a-476b-b7e1-76bb4c6ef18d-config-volume" (OuterVolumeSpecName: "config-volume") pod "f211bc8c-978a-476b-b7e1-76bb4c6ef18d" (UID: "f211bc8c-978a-476b-b7e1-76bb4c6ef18d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 05:30:03 crc kubenswrapper[4799]: I1129 05:30:03.591341 4799 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f211bc8c-978a-476b-b7e1-76bb4c6ef18d-config-volume\") on node \"crc\" DevicePath \"\"" Nov 29 05:30:03 crc kubenswrapper[4799]: I1129 05:30:03.598607 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f211bc8c-978a-476b-b7e1-76bb4c6ef18d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f211bc8c-978a-476b-b7e1-76bb4c6ef18d" (UID: "f211bc8c-978a-476b-b7e1-76bb4c6ef18d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:30:03 crc kubenswrapper[4799]: I1129 05:30:03.598773 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f211bc8c-978a-476b-b7e1-76bb4c6ef18d-kube-api-access-ktrcg" (OuterVolumeSpecName: "kube-api-access-ktrcg") pod "f211bc8c-978a-476b-b7e1-76bb4c6ef18d" (UID: "f211bc8c-978a-476b-b7e1-76bb4c6ef18d"). InnerVolumeSpecName "kube-api-access-ktrcg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:30:03 crc kubenswrapper[4799]: I1129 05:30:03.692913 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ktrcg\" (UniqueName: \"kubernetes.io/projected/f211bc8c-978a-476b-b7e1-76bb4c6ef18d-kube-api-access-ktrcg\") on node \"crc\" DevicePath \"\"" Nov 29 05:30:03 crc kubenswrapper[4799]: I1129 05:30:03.692951 4799 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f211bc8c-978a-476b-b7e1-76bb4c6ef18d-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 29 05:30:04 crc kubenswrapper[4799]: I1129 05:30:04.183379 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29406570-r4g4v" event={"ID":"f211bc8c-978a-476b-b7e1-76bb4c6ef18d","Type":"ContainerDied","Data":"06b837a5e377bd20f5c2b72a1996b39470d3da1ca29862a77e2e4afb319632dd"} Nov 29 05:30:04 crc kubenswrapper[4799]: I1129 05:30:04.183450 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="06b837a5e377bd20f5c2b72a1996b39470d3da1ca29862a77e2e4afb319632dd" Nov 29 05:30:04 crc kubenswrapper[4799]: I1129 05:30:04.183538 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406570-r4g4v" Nov 29 05:30:04 crc kubenswrapper[4799]: I1129 05:30:04.636714 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406525-md8rr"] Nov 29 05:30:04 crc kubenswrapper[4799]: I1129 05:30:04.646820 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406525-md8rr"] Nov 29 05:30:04 crc kubenswrapper[4799]: I1129 05:30:04.674027 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c614b62c-cd99-4f49-8939-25fe713e4809" path="/var/lib/kubelet/pods/c614b62c-cd99-4f49-8939-25fe713e4809/volumes" Nov 29 05:30:06 crc kubenswrapper[4799]: I1129 05:30:06.987975 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Nov 29 05:30:10 crc kubenswrapper[4799]: I1129 05:30:10.202196 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Nov 29 05:30:10 crc kubenswrapper[4799]: I1129 05:30:10.233594 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 29 05:30:14 crc kubenswrapper[4799]: I1129 05:30:14.670654 4799 scope.go:117] "RemoveContainer" containerID="e6d4b3d1d5b5e1eeb13508d67de3d3c3aac52e76337b55e53793185c0d001d82" Nov 29 05:30:14 crc kubenswrapper[4799]: E1129 05:30:14.672142 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:30:26 crc kubenswrapper[4799]: I1129 05:30:26.660936 4799 scope.go:117] "RemoveContainer" containerID="e6d4b3d1d5b5e1eeb13508d67de3d3c3aac52e76337b55e53793185c0d001d82" Nov 29 05:30:26 crc kubenswrapper[4799]: E1129 05:30:26.661878 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:30:39 crc kubenswrapper[4799]: I1129 05:30:39.660746 4799 scope.go:117] "RemoveContainer" containerID="e6d4b3d1d5b5e1eeb13508d67de3d3c3aac52e76337b55e53793185c0d001d82" Nov 29 05:30:39 crc kubenswrapper[4799]: E1129 05:30:39.662919 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:30:50 crc kubenswrapper[4799]: I1129 05:30:50.662718 4799 scope.go:117] "RemoveContainer" containerID="e6d4b3d1d5b5e1eeb13508d67de3d3c3aac52e76337b55e53793185c0d001d82" Nov 29 05:30:50 crc kubenswrapper[4799]: E1129 05:30:50.663438 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:30:58 crc kubenswrapper[4799]: I1129 05:30:58.119512 4799 scope.go:117] "RemoveContainer" containerID="2abf20d02fcbf7ac03d4005350466966763796841a92dd495d6066c8ab273a23" Nov 29 05:31:05 crc kubenswrapper[4799]: I1129 05:31:05.659633 4799 scope.go:117] "RemoveContainer" containerID="e6d4b3d1d5b5e1eeb13508d67de3d3c3aac52e76337b55e53793185c0d001d82" Nov 29 05:31:05 crc kubenswrapper[4799]: E1129 05:31:05.660642 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:31:10 crc kubenswrapper[4799]: I1129 05:31:10.804921 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Nov 29 05:31:10 crc kubenswrapper[4799]: E1129 05:31:10.806904 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f09087c5-87d5-4cd1-9a91-cbe48713d9e9" containerName="horizon" Nov 29 05:31:10 crc kubenswrapper[4799]: I1129 05:31:10.806929 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f09087c5-87d5-4cd1-9a91-cbe48713d9e9" containerName="horizon" Nov 29 05:31:10 crc kubenswrapper[4799]: E1129 05:31:10.806962 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f211bc8c-978a-476b-b7e1-76bb4c6ef18d" containerName="collect-profiles" Nov 29 05:31:10 crc kubenswrapper[4799]: I1129 05:31:10.806974 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f211bc8c-978a-476b-b7e1-76bb4c6ef18d" containerName="collect-profiles" Nov 29 05:31:10 crc kubenswrapper[4799]: E1129 05:31:10.807008 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f09087c5-87d5-4cd1-9a91-cbe48713d9e9" containerName="horizon-log" Nov 29 05:31:10 crc kubenswrapper[4799]: I1129 05:31:10.807019 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f09087c5-87d5-4cd1-9a91-cbe48713d9e9" containerName="horizon-log" Nov 29 05:31:10 crc kubenswrapper[4799]: I1129 05:31:10.807382 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="f09087c5-87d5-4cd1-9a91-cbe48713d9e9" containerName="horizon-log" Nov 29 05:31:10 crc kubenswrapper[4799]: I1129 05:31:10.807423 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="f09087c5-87d5-4cd1-9a91-cbe48713d9e9" containerName="horizon" Nov 29 05:31:10 crc kubenswrapper[4799]: I1129 05:31:10.807456 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="f211bc8c-978a-476b-b7e1-76bb4c6ef18d" containerName="collect-profiles" Nov 29 05:31:10 crc kubenswrapper[4799]: I1129 05:31:10.808712 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 29 05:31:10 crc kubenswrapper[4799]: I1129 05:31:10.811711 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Nov 29 05:31:10 crc kubenswrapper[4799]: I1129 05:31:10.813169 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-n9z7f" Nov 29 05:31:10 crc kubenswrapper[4799]: I1129 05:31:10.813294 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Nov 29 05:31:10 crc kubenswrapper[4799]: I1129 05:31:10.814302 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Nov 29 05:31:10 crc kubenswrapper[4799]: I1129 05:31:10.832167 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Nov 29 05:31:10 crc kubenswrapper[4799]: I1129 05:31:10.988294 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fs2z\" (UniqueName: \"kubernetes.io/projected/c4a30984-7c0c-493b-b752-b603232c837f-kube-api-access-2fs2z\") pod \"tempest-tests-tempest\" (UID: \"c4a30984-7c0c-493b-b752-b603232c837f\") " pod="openstack/tempest-tests-tempest" Nov 29 05:31:10 crc kubenswrapper[4799]: I1129 05:31:10.988403 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c4a30984-7c0c-493b-b752-b603232c837f-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"c4a30984-7c0c-493b-b752-b603232c837f\") " pod="openstack/tempest-tests-tempest" Nov 29 05:31:10 crc kubenswrapper[4799]: I1129 05:31:10.988451 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"tempest-tests-tempest\" (UID: \"c4a30984-7c0c-493b-b752-b603232c837f\") " pod="openstack/tempest-tests-tempest" Nov 29 05:31:10 crc kubenswrapper[4799]: I1129 05:31:10.988490 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/c4a30984-7c0c-493b-b752-b603232c837f-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"c4a30984-7c0c-493b-b752-b603232c837f\") " pod="openstack/tempest-tests-tempest" Nov 29 05:31:10 crc kubenswrapper[4799]: I1129 05:31:10.988555 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/c4a30984-7c0c-493b-b752-b603232c837f-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"c4a30984-7c0c-493b-b752-b603232c837f\") " pod="openstack/tempest-tests-tempest" Nov 29 05:31:10 crc kubenswrapper[4799]: I1129 05:31:10.988649 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c4a30984-7c0c-493b-b752-b603232c837f-config-data\") pod \"tempest-tests-tempest\" (UID: \"c4a30984-7c0c-493b-b752-b603232c837f\") " pod="openstack/tempest-tests-tempest" Nov 29 05:31:10 crc kubenswrapper[4799]: I1129 05:31:10.989083 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c4a30984-7c0c-493b-b752-b603232c837f-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"c4a30984-7c0c-493b-b752-b603232c837f\") " pod="openstack/tempest-tests-tempest" Nov 29 05:31:10 crc kubenswrapper[4799]: I1129 05:31:10.989232 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c4a30984-7c0c-493b-b752-b603232c837f-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"c4a30984-7c0c-493b-b752-b603232c837f\") " pod="openstack/tempest-tests-tempest" Nov 29 05:31:10 crc kubenswrapper[4799]: I1129 05:31:10.989273 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/c4a30984-7c0c-493b-b752-b603232c837f-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"c4a30984-7c0c-493b-b752-b603232c837f\") " pod="openstack/tempest-tests-tempest" Nov 29 05:31:11 crc kubenswrapper[4799]: I1129 05:31:11.091272 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c4a30984-7c0c-493b-b752-b603232c837f-config-data\") pod \"tempest-tests-tempest\" (UID: \"c4a30984-7c0c-493b-b752-b603232c837f\") " pod="openstack/tempest-tests-tempest" Nov 29 05:31:11 crc kubenswrapper[4799]: I1129 05:31:11.091353 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c4a30984-7c0c-493b-b752-b603232c837f-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"c4a30984-7c0c-493b-b752-b603232c837f\") " pod="openstack/tempest-tests-tempest" Nov 29 05:31:11 crc kubenswrapper[4799]: I1129 05:31:11.091414 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c4a30984-7c0c-493b-b752-b603232c837f-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"c4a30984-7c0c-493b-b752-b603232c837f\") " pod="openstack/tempest-tests-tempest" Nov 29 05:31:11 crc kubenswrapper[4799]: I1129 05:31:11.091450 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/c4a30984-7c0c-493b-b752-b603232c837f-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"c4a30984-7c0c-493b-b752-b603232c837f\") " pod="openstack/tempest-tests-tempest" Nov 29 05:31:11 crc kubenswrapper[4799]: I1129 05:31:11.091502 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fs2z\" (UniqueName: \"kubernetes.io/projected/c4a30984-7c0c-493b-b752-b603232c837f-kube-api-access-2fs2z\") pod \"tempest-tests-tempest\" (UID: \"c4a30984-7c0c-493b-b752-b603232c837f\") " pod="openstack/tempest-tests-tempest" Nov 29 05:31:11 crc kubenswrapper[4799]: I1129 05:31:11.091545 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c4a30984-7c0c-493b-b752-b603232c837f-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"c4a30984-7c0c-493b-b752-b603232c837f\") " pod="openstack/tempest-tests-tempest" Nov 29 05:31:11 crc kubenswrapper[4799]: I1129 05:31:11.091578 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"tempest-tests-tempest\" (UID: \"c4a30984-7c0c-493b-b752-b603232c837f\") " pod="openstack/tempest-tests-tempest" Nov 29 05:31:11 crc kubenswrapper[4799]: I1129 05:31:11.091929 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/c4a30984-7c0c-493b-b752-b603232c837f-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"c4a30984-7c0c-493b-b752-b603232c837f\") " pod="openstack/tempest-tests-tempest" Nov 29 05:31:11 crc kubenswrapper[4799]: I1129 05:31:11.092090 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"tempest-tests-tempest\" (UID: \"c4a30984-7c0c-493b-b752-b603232c837f\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/tempest-tests-tempest" Nov 29 05:31:11 crc kubenswrapper[4799]: I1129 05:31:11.092383 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/c4a30984-7c0c-493b-b752-b603232c837f-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"c4a30984-7c0c-493b-b752-b603232c837f\") " pod="openstack/tempest-tests-tempest" Nov 29 05:31:11 crc kubenswrapper[4799]: I1129 05:31:11.092384 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/c4a30984-7c0c-493b-b752-b603232c837f-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"c4a30984-7c0c-493b-b752-b603232c837f\") " pod="openstack/tempest-tests-tempest" Nov 29 05:31:11 crc kubenswrapper[4799]: I1129 05:31:11.093361 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c4a30984-7c0c-493b-b752-b603232c837f-config-data\") pod \"tempest-tests-tempest\" (UID: \"c4a30984-7c0c-493b-b752-b603232c837f\") " pod="openstack/tempest-tests-tempest" Nov 29 05:31:11 crc kubenswrapper[4799]: I1129 05:31:11.093916 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/c4a30984-7c0c-493b-b752-b603232c837f-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"c4a30984-7c0c-493b-b752-b603232c837f\") " pod="openstack/tempest-tests-tempest" Nov 29 05:31:11 crc kubenswrapper[4799]: I1129 05:31:11.094292 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c4a30984-7c0c-493b-b752-b603232c837f-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"c4a30984-7c0c-493b-b752-b603232c837f\") " pod="openstack/tempest-tests-tempest" Nov 29 05:31:11 crc kubenswrapper[4799]: I1129 05:31:11.098426 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/c4a30984-7c0c-493b-b752-b603232c837f-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"c4a30984-7c0c-493b-b752-b603232c837f\") " pod="openstack/tempest-tests-tempest" Nov 29 05:31:11 crc kubenswrapper[4799]: I1129 05:31:11.098864 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c4a30984-7c0c-493b-b752-b603232c837f-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"c4a30984-7c0c-493b-b752-b603232c837f\") " pod="openstack/tempest-tests-tempest" Nov 29 05:31:11 crc kubenswrapper[4799]: I1129 05:31:11.098997 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c4a30984-7c0c-493b-b752-b603232c837f-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"c4a30984-7c0c-493b-b752-b603232c837f\") " pod="openstack/tempest-tests-tempest" Nov 29 05:31:11 crc kubenswrapper[4799]: I1129 05:31:11.107961 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fs2z\" (UniqueName: \"kubernetes.io/projected/c4a30984-7c0c-493b-b752-b603232c837f-kube-api-access-2fs2z\") pod \"tempest-tests-tempest\" (UID: \"c4a30984-7c0c-493b-b752-b603232c837f\") " pod="openstack/tempest-tests-tempest" Nov 29 05:31:11 crc kubenswrapper[4799]: I1129 05:31:11.127662 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"tempest-tests-tempest\" (UID: \"c4a30984-7c0c-493b-b752-b603232c837f\") " pod="openstack/tempest-tests-tempest" Nov 29 05:31:11 crc kubenswrapper[4799]: I1129 05:31:11.151509 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 29 05:31:11 crc kubenswrapper[4799]: I1129 05:31:11.662606 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 29 05:31:11 crc kubenswrapper[4799]: I1129 05:31:11.665508 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Nov 29 05:31:12 crc kubenswrapper[4799]: I1129 05:31:12.095900 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"c4a30984-7c0c-493b-b752-b603232c837f","Type":"ContainerStarted","Data":"614a0889570a9c038cf129ec0198f707974bf7e8c1fc2a5e432330ea1f334508"} Nov 29 05:31:20 crc kubenswrapper[4799]: I1129 05:31:20.659509 4799 scope.go:117] "RemoveContainer" containerID="e6d4b3d1d5b5e1eeb13508d67de3d3c3aac52e76337b55e53793185c0d001d82" Nov 29 05:31:20 crc kubenswrapper[4799]: E1129 05:31:20.661967 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:31:31 crc kubenswrapper[4799]: I1129 05:31:31.660386 4799 scope.go:117] "RemoveContainer" containerID="e6d4b3d1d5b5e1eeb13508d67de3d3c3aac52e76337b55e53793185c0d001d82" Nov 29 05:31:31 crc kubenswrapper[4799]: E1129 05:31:31.661680 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:31:45 crc kubenswrapper[4799]: I1129 05:31:45.660496 4799 scope.go:117] "RemoveContainer" containerID="e6d4b3d1d5b5e1eeb13508d67de3d3c3aac52e76337b55e53793185c0d001d82" Nov 29 05:31:45 crc kubenswrapper[4799]: E1129 05:31:45.661389 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:31:46 crc kubenswrapper[4799]: E1129 05:31:46.304769 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Nov 29 05:31:46 crc kubenswrapper[4799]: E1129 05:31:46.305482 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2fs2z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(c4a30984-7c0c-493b-b752-b603232c837f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 29 05:31:46 crc kubenswrapper[4799]: E1129 05:31:46.306975 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="c4a30984-7c0c-493b-b752-b603232c837f" Nov 29 05:31:46 crc kubenswrapper[4799]: E1129 05:31:46.535678 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="c4a30984-7c0c-493b-b752-b603232c837f" Nov 29 05:31:46 crc kubenswrapper[4799]: I1129 05:31:46.958219 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lc4k5"] Nov 29 05:31:46 crc kubenswrapper[4799]: I1129 05:31:46.961655 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lc4k5" Nov 29 05:31:46 crc kubenswrapper[4799]: I1129 05:31:46.972253 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lc4k5"] Nov 29 05:31:47 crc kubenswrapper[4799]: I1129 05:31:47.109604 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/520ff9ee-edbe-4684-925a-c220002c5cc4-catalog-content\") pod \"certified-operators-lc4k5\" (UID: \"520ff9ee-edbe-4684-925a-c220002c5cc4\") " pod="openshift-marketplace/certified-operators-lc4k5" Nov 29 05:31:47 crc kubenswrapper[4799]: I1129 05:31:47.109719 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/520ff9ee-edbe-4684-925a-c220002c5cc4-utilities\") pod \"certified-operators-lc4k5\" (UID: \"520ff9ee-edbe-4684-925a-c220002c5cc4\") " pod="openshift-marketplace/certified-operators-lc4k5" Nov 29 05:31:47 crc kubenswrapper[4799]: I1129 05:31:47.110172 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czh28\" (UniqueName: \"kubernetes.io/projected/520ff9ee-edbe-4684-925a-c220002c5cc4-kube-api-access-czh28\") pod \"certified-operators-lc4k5\" (UID: \"520ff9ee-edbe-4684-925a-c220002c5cc4\") " pod="openshift-marketplace/certified-operators-lc4k5" Nov 29 05:31:47 crc kubenswrapper[4799]: I1129 05:31:47.212991 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/520ff9ee-edbe-4684-925a-c220002c5cc4-catalog-content\") pod \"certified-operators-lc4k5\" (UID: \"520ff9ee-edbe-4684-925a-c220002c5cc4\") " pod="openshift-marketplace/certified-operators-lc4k5" Nov 29 05:31:47 crc kubenswrapper[4799]: I1129 05:31:47.213082 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/520ff9ee-edbe-4684-925a-c220002c5cc4-utilities\") pod \"certified-operators-lc4k5\" (UID: \"520ff9ee-edbe-4684-925a-c220002c5cc4\") " pod="openshift-marketplace/certified-operators-lc4k5" Nov 29 05:31:47 crc kubenswrapper[4799]: I1129 05:31:47.213137 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czh28\" (UniqueName: \"kubernetes.io/projected/520ff9ee-edbe-4684-925a-c220002c5cc4-kube-api-access-czh28\") pod \"certified-operators-lc4k5\" (UID: \"520ff9ee-edbe-4684-925a-c220002c5cc4\") " pod="openshift-marketplace/certified-operators-lc4k5" Nov 29 05:31:47 crc kubenswrapper[4799]: I1129 05:31:47.213636 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/520ff9ee-edbe-4684-925a-c220002c5cc4-catalog-content\") pod \"certified-operators-lc4k5\" (UID: \"520ff9ee-edbe-4684-925a-c220002c5cc4\") " pod="openshift-marketplace/certified-operators-lc4k5" Nov 29 05:31:47 crc kubenswrapper[4799]: I1129 05:31:47.213673 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/520ff9ee-edbe-4684-925a-c220002c5cc4-utilities\") pod \"certified-operators-lc4k5\" (UID: \"520ff9ee-edbe-4684-925a-c220002c5cc4\") " pod="openshift-marketplace/certified-operators-lc4k5" Nov 29 05:31:47 crc kubenswrapper[4799]: I1129 05:31:47.241342 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czh28\" (UniqueName: \"kubernetes.io/projected/520ff9ee-edbe-4684-925a-c220002c5cc4-kube-api-access-czh28\") pod \"certified-operators-lc4k5\" (UID: \"520ff9ee-edbe-4684-925a-c220002c5cc4\") " pod="openshift-marketplace/certified-operators-lc4k5" Nov 29 05:31:47 crc kubenswrapper[4799]: I1129 05:31:47.307012 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lc4k5" Nov 29 05:31:47 crc kubenswrapper[4799]: I1129 05:31:47.881976 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lc4k5"] Nov 29 05:31:48 crc kubenswrapper[4799]: I1129 05:31:48.559297 4799 generic.go:334] "Generic (PLEG): container finished" podID="520ff9ee-edbe-4684-925a-c220002c5cc4" containerID="6f05c112bd93fd1948fd923d5ac9c763942d7af20ae5056180f11ae141288832" exitCode=0 Nov 29 05:31:48 crc kubenswrapper[4799]: I1129 05:31:48.560111 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lc4k5" event={"ID":"520ff9ee-edbe-4684-925a-c220002c5cc4","Type":"ContainerDied","Data":"6f05c112bd93fd1948fd923d5ac9c763942d7af20ae5056180f11ae141288832"} Nov 29 05:31:48 crc kubenswrapper[4799]: I1129 05:31:48.561038 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lc4k5" event={"ID":"520ff9ee-edbe-4684-925a-c220002c5cc4","Type":"ContainerStarted","Data":"36af05e6414874ff158fb78b5bae84b0136603a3bb029eb2ca9b1c884ed85f5f"} Nov 29 05:31:49 crc kubenswrapper[4799]: I1129 05:31:49.576094 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lc4k5" event={"ID":"520ff9ee-edbe-4684-925a-c220002c5cc4","Type":"ContainerStarted","Data":"b815039f205ccbefdb1c4b6dff6fb4970bd0f795ac31ee73beb034cea396124a"} Nov 29 05:31:50 crc kubenswrapper[4799]: I1129 05:31:50.340036 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zf798"] Nov 29 05:31:50 crc kubenswrapper[4799]: I1129 05:31:50.342372 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zf798" Nov 29 05:31:50 crc kubenswrapper[4799]: I1129 05:31:50.439755 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zf798"] Nov 29 05:31:50 crc kubenswrapper[4799]: I1129 05:31:50.511838 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/723bd6ac-25ca-405b-8418-4e60fa6e2d22-catalog-content\") pod \"community-operators-zf798\" (UID: \"723bd6ac-25ca-405b-8418-4e60fa6e2d22\") " pod="openshift-marketplace/community-operators-zf798" Nov 29 05:31:50 crc kubenswrapper[4799]: I1129 05:31:50.511916 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/723bd6ac-25ca-405b-8418-4e60fa6e2d22-utilities\") pod \"community-operators-zf798\" (UID: \"723bd6ac-25ca-405b-8418-4e60fa6e2d22\") " pod="openshift-marketplace/community-operators-zf798" Nov 29 05:31:50 crc kubenswrapper[4799]: I1129 05:31:50.511972 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-849st\" (UniqueName: \"kubernetes.io/projected/723bd6ac-25ca-405b-8418-4e60fa6e2d22-kube-api-access-849st\") pod \"community-operators-zf798\" (UID: \"723bd6ac-25ca-405b-8418-4e60fa6e2d22\") " pod="openshift-marketplace/community-operators-zf798" Nov 29 05:31:50 crc kubenswrapper[4799]: I1129 05:31:50.587562 4799 generic.go:334] "Generic (PLEG): container finished" podID="520ff9ee-edbe-4684-925a-c220002c5cc4" containerID="b815039f205ccbefdb1c4b6dff6fb4970bd0f795ac31ee73beb034cea396124a" exitCode=0 Nov 29 05:31:50 crc kubenswrapper[4799]: I1129 05:31:50.587631 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lc4k5" event={"ID":"520ff9ee-edbe-4684-925a-c220002c5cc4","Type":"ContainerDied","Data":"b815039f205ccbefdb1c4b6dff6fb4970bd0f795ac31ee73beb034cea396124a"} Nov 29 05:31:50 crc kubenswrapper[4799]: I1129 05:31:50.613778 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/723bd6ac-25ca-405b-8418-4e60fa6e2d22-catalog-content\") pod \"community-operators-zf798\" (UID: \"723bd6ac-25ca-405b-8418-4e60fa6e2d22\") " pod="openshift-marketplace/community-operators-zf798" Nov 29 05:31:50 crc kubenswrapper[4799]: I1129 05:31:50.613889 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/723bd6ac-25ca-405b-8418-4e60fa6e2d22-utilities\") pod \"community-operators-zf798\" (UID: \"723bd6ac-25ca-405b-8418-4e60fa6e2d22\") " pod="openshift-marketplace/community-operators-zf798" Nov 29 05:31:50 crc kubenswrapper[4799]: I1129 05:31:50.613950 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-849st\" (UniqueName: \"kubernetes.io/projected/723bd6ac-25ca-405b-8418-4e60fa6e2d22-kube-api-access-849st\") pod \"community-operators-zf798\" (UID: \"723bd6ac-25ca-405b-8418-4e60fa6e2d22\") " pod="openshift-marketplace/community-operators-zf798" Nov 29 05:31:50 crc kubenswrapper[4799]: I1129 05:31:50.614556 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/723bd6ac-25ca-405b-8418-4e60fa6e2d22-catalog-content\") pod \"community-operators-zf798\" (UID: \"723bd6ac-25ca-405b-8418-4e60fa6e2d22\") " pod="openshift-marketplace/community-operators-zf798" Nov 29 05:31:50 crc kubenswrapper[4799]: I1129 05:31:50.614582 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/723bd6ac-25ca-405b-8418-4e60fa6e2d22-utilities\") pod \"community-operators-zf798\" (UID: \"723bd6ac-25ca-405b-8418-4e60fa6e2d22\") " pod="openshift-marketplace/community-operators-zf798" Nov 29 05:31:50 crc kubenswrapper[4799]: I1129 05:31:50.643084 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-849st\" (UniqueName: \"kubernetes.io/projected/723bd6ac-25ca-405b-8418-4e60fa6e2d22-kube-api-access-849st\") pod \"community-operators-zf798\" (UID: \"723bd6ac-25ca-405b-8418-4e60fa6e2d22\") " pod="openshift-marketplace/community-operators-zf798" Nov 29 05:31:50 crc kubenswrapper[4799]: I1129 05:31:50.682920 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zf798" Nov 29 05:31:51 crc kubenswrapper[4799]: I1129 05:31:51.238367 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zf798"] Nov 29 05:31:51 crc kubenswrapper[4799]: W1129 05:31:51.241626 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod723bd6ac_25ca_405b_8418_4e60fa6e2d22.slice/crio-dcdddb044f12b079026688e77e4269b1b77fc79c6a6a383b28d23c65ee1fb7a2 WatchSource:0}: Error finding container dcdddb044f12b079026688e77e4269b1b77fc79c6a6a383b28d23c65ee1fb7a2: Status 404 returned error can't find the container with id dcdddb044f12b079026688e77e4269b1b77fc79c6a6a383b28d23c65ee1fb7a2 Nov 29 05:31:51 crc kubenswrapper[4799]: I1129 05:31:51.606181 4799 generic.go:334] "Generic (PLEG): container finished" podID="723bd6ac-25ca-405b-8418-4e60fa6e2d22" containerID="f64548328f3fa19d3b235dcc7f5bbda8dc20742f77cac05b3e87a2bd8bb4b9a7" exitCode=0 Nov 29 05:31:51 crc kubenswrapper[4799]: I1129 05:31:51.606248 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zf798" event={"ID":"723bd6ac-25ca-405b-8418-4e60fa6e2d22","Type":"ContainerDied","Data":"f64548328f3fa19d3b235dcc7f5bbda8dc20742f77cac05b3e87a2bd8bb4b9a7"} Nov 29 05:31:51 crc kubenswrapper[4799]: I1129 05:31:51.606688 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zf798" event={"ID":"723bd6ac-25ca-405b-8418-4e60fa6e2d22","Type":"ContainerStarted","Data":"dcdddb044f12b079026688e77e4269b1b77fc79c6a6a383b28d23c65ee1fb7a2"} Nov 29 05:31:51 crc kubenswrapper[4799]: I1129 05:31:51.613811 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lc4k5" event={"ID":"520ff9ee-edbe-4684-925a-c220002c5cc4","Type":"ContainerStarted","Data":"84e51c8615ea994610df768cbbb5ecbb2835020efb9a2bca864e1455ebfb7b2c"} Nov 29 05:31:52 crc kubenswrapper[4799]: I1129 05:31:52.630472 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zf798" event={"ID":"723bd6ac-25ca-405b-8418-4e60fa6e2d22","Type":"ContainerStarted","Data":"47737218625d4b17dfa86e1e0e19f541fc2c97683e95b025f1667bf14de2d108"} Nov 29 05:31:52 crc kubenswrapper[4799]: I1129 05:31:52.664833 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lc4k5" podStartSLOduration=3.9199938640000003 podStartE2EDuration="6.664803712s" podCreationTimestamp="2025-11-29 05:31:46 +0000 UTC" firstStartedPulling="2025-11-29 05:31:48.562014039 +0000 UTC m=+3184.204944449" lastFinishedPulling="2025-11-29 05:31:51.306823897 +0000 UTC m=+3186.949754297" observedRunningTime="2025-11-29 05:31:51.657628628 +0000 UTC m=+3187.300559078" watchObservedRunningTime="2025-11-29 05:31:52.664803712 +0000 UTC m=+3188.307734122" Nov 29 05:31:53 crc kubenswrapper[4799]: I1129 05:31:53.646490 4799 generic.go:334] "Generic (PLEG): container finished" podID="723bd6ac-25ca-405b-8418-4e60fa6e2d22" containerID="47737218625d4b17dfa86e1e0e19f541fc2c97683e95b025f1667bf14de2d108" exitCode=0 Nov 29 05:31:53 crc kubenswrapper[4799]: I1129 05:31:53.646544 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zf798" event={"ID":"723bd6ac-25ca-405b-8418-4e60fa6e2d22","Type":"ContainerDied","Data":"47737218625d4b17dfa86e1e0e19f541fc2c97683e95b025f1667bf14de2d108"} Nov 29 05:31:55 crc kubenswrapper[4799]: I1129 05:31:55.669646 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zf798" event={"ID":"723bd6ac-25ca-405b-8418-4e60fa6e2d22","Type":"ContainerStarted","Data":"aa6afffe3039808fe86eaf10efc4620f3db866998c3f032070427e362de2cc3b"} Nov 29 05:31:55 crc kubenswrapper[4799]: I1129 05:31:55.696074 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zf798" podStartSLOduration=3.120489365 podStartE2EDuration="5.696055176s" podCreationTimestamp="2025-11-29 05:31:50 +0000 UTC" firstStartedPulling="2025-11-29 05:31:51.608443864 +0000 UTC m=+3187.251374264" lastFinishedPulling="2025-11-29 05:31:54.184009665 +0000 UTC m=+3189.826940075" observedRunningTime="2025-11-29 05:31:55.691684632 +0000 UTC m=+3191.334615102" watchObservedRunningTime="2025-11-29 05:31:55.696055176 +0000 UTC m=+3191.338985576" Nov 29 05:31:56 crc kubenswrapper[4799]: I1129 05:31:56.724322 4799 scope.go:117] "RemoveContainer" containerID="e6d4b3d1d5b5e1eeb13508d67de3d3c3aac52e76337b55e53793185c0d001d82" Nov 29 05:31:56 crc kubenswrapper[4799]: E1129 05:31:56.724670 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:31:57 crc kubenswrapper[4799]: I1129 05:31:57.308315 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lc4k5" Nov 29 05:31:57 crc kubenswrapper[4799]: I1129 05:31:57.308799 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lc4k5" Nov 29 05:31:57 crc kubenswrapper[4799]: I1129 05:31:57.356808 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lc4k5" Nov 29 05:31:57 crc kubenswrapper[4799]: I1129 05:31:57.829309 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lc4k5" Nov 29 05:31:58 crc kubenswrapper[4799]: I1129 05:31:58.539210 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lc4k5"] Nov 29 05:31:59 crc kubenswrapper[4799]: I1129 05:31:59.784409 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-lc4k5" podUID="520ff9ee-edbe-4684-925a-c220002c5cc4" containerName="registry-server" containerID="cri-o://84e51c8615ea994610df768cbbb5ecbb2835020efb9a2bca864e1455ebfb7b2c" gracePeriod=2 Nov 29 05:32:00 crc kubenswrapper[4799]: I1129 05:32:00.227977 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lc4k5" Nov 29 05:32:00 crc kubenswrapper[4799]: I1129 05:32:00.327172 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-czh28\" (UniqueName: \"kubernetes.io/projected/520ff9ee-edbe-4684-925a-c220002c5cc4-kube-api-access-czh28\") pod \"520ff9ee-edbe-4684-925a-c220002c5cc4\" (UID: \"520ff9ee-edbe-4684-925a-c220002c5cc4\") " Nov 29 05:32:00 crc kubenswrapper[4799]: I1129 05:32:00.327390 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/520ff9ee-edbe-4684-925a-c220002c5cc4-utilities\") pod \"520ff9ee-edbe-4684-925a-c220002c5cc4\" (UID: \"520ff9ee-edbe-4684-925a-c220002c5cc4\") " Nov 29 05:32:00 crc kubenswrapper[4799]: I1129 05:32:00.327533 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/520ff9ee-edbe-4684-925a-c220002c5cc4-catalog-content\") pod \"520ff9ee-edbe-4684-925a-c220002c5cc4\" (UID: \"520ff9ee-edbe-4684-925a-c220002c5cc4\") " Nov 29 05:32:00 crc kubenswrapper[4799]: I1129 05:32:00.329119 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/520ff9ee-edbe-4684-925a-c220002c5cc4-utilities" (OuterVolumeSpecName: "utilities") pod "520ff9ee-edbe-4684-925a-c220002c5cc4" (UID: "520ff9ee-edbe-4684-925a-c220002c5cc4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:32:00 crc kubenswrapper[4799]: I1129 05:32:00.337168 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/520ff9ee-edbe-4684-925a-c220002c5cc4-kube-api-access-czh28" (OuterVolumeSpecName: "kube-api-access-czh28") pod "520ff9ee-edbe-4684-925a-c220002c5cc4" (UID: "520ff9ee-edbe-4684-925a-c220002c5cc4"). InnerVolumeSpecName "kube-api-access-czh28". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:32:00 crc kubenswrapper[4799]: I1129 05:32:00.399142 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/520ff9ee-edbe-4684-925a-c220002c5cc4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "520ff9ee-edbe-4684-925a-c220002c5cc4" (UID: "520ff9ee-edbe-4684-925a-c220002c5cc4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:32:00 crc kubenswrapper[4799]: I1129 05:32:00.431205 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/520ff9ee-edbe-4684-925a-c220002c5cc4-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 05:32:00 crc kubenswrapper[4799]: I1129 05:32:00.431265 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/520ff9ee-edbe-4684-925a-c220002c5cc4-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 05:32:00 crc kubenswrapper[4799]: I1129 05:32:00.431279 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-czh28\" (UniqueName: \"kubernetes.io/projected/520ff9ee-edbe-4684-925a-c220002c5cc4-kube-api-access-czh28\") on node \"crc\" DevicePath \"\"" Nov 29 05:32:00 crc kubenswrapper[4799]: I1129 05:32:00.683914 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zf798" Nov 29 05:32:00 crc kubenswrapper[4799]: I1129 05:32:00.684358 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zf798" Nov 29 05:32:00 crc kubenswrapper[4799]: I1129 05:32:00.732464 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zf798" Nov 29 05:32:00 crc kubenswrapper[4799]: I1129 05:32:00.797605 4799 generic.go:334] "Generic (PLEG): container finished" podID="520ff9ee-edbe-4684-925a-c220002c5cc4" containerID="84e51c8615ea994610df768cbbb5ecbb2835020efb9a2bca864e1455ebfb7b2c" exitCode=0 Nov 29 05:32:00 crc kubenswrapper[4799]: I1129 05:32:00.797681 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lc4k5" Nov 29 05:32:00 crc kubenswrapper[4799]: I1129 05:32:00.797707 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lc4k5" event={"ID":"520ff9ee-edbe-4684-925a-c220002c5cc4","Type":"ContainerDied","Data":"84e51c8615ea994610df768cbbb5ecbb2835020efb9a2bca864e1455ebfb7b2c"} Nov 29 05:32:00 crc kubenswrapper[4799]: I1129 05:32:00.798949 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lc4k5" event={"ID":"520ff9ee-edbe-4684-925a-c220002c5cc4","Type":"ContainerDied","Data":"36af05e6414874ff158fb78b5bae84b0136603a3bb029eb2ca9b1c884ed85f5f"} Nov 29 05:32:00 crc kubenswrapper[4799]: I1129 05:32:00.798970 4799 scope.go:117] "RemoveContainer" containerID="84e51c8615ea994610df768cbbb5ecbb2835020efb9a2bca864e1455ebfb7b2c" Nov 29 05:32:00 crc kubenswrapper[4799]: I1129 05:32:00.832464 4799 scope.go:117] "RemoveContainer" containerID="b815039f205ccbefdb1c4b6dff6fb4970bd0f795ac31ee73beb034cea396124a" Nov 29 05:32:00 crc kubenswrapper[4799]: I1129 05:32:00.846948 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lc4k5"] Nov 29 05:32:00 crc kubenswrapper[4799]: I1129 05:32:00.859979 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zf798" Nov 29 05:32:00 crc kubenswrapper[4799]: I1129 05:32:00.862617 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-lc4k5"] Nov 29 05:32:00 crc kubenswrapper[4799]: I1129 05:32:00.895222 4799 scope.go:117] "RemoveContainer" containerID="6f05c112bd93fd1948fd923d5ac9c763942d7af20ae5056180f11ae141288832" Nov 29 05:32:00 crc kubenswrapper[4799]: I1129 05:32:00.928183 4799 scope.go:117] "RemoveContainer" containerID="84e51c8615ea994610df768cbbb5ecbb2835020efb9a2bca864e1455ebfb7b2c" Nov 29 05:32:00 crc kubenswrapper[4799]: E1129 05:32:00.928658 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84e51c8615ea994610df768cbbb5ecbb2835020efb9a2bca864e1455ebfb7b2c\": container with ID starting with 84e51c8615ea994610df768cbbb5ecbb2835020efb9a2bca864e1455ebfb7b2c not found: ID does not exist" containerID="84e51c8615ea994610df768cbbb5ecbb2835020efb9a2bca864e1455ebfb7b2c" Nov 29 05:32:00 crc kubenswrapper[4799]: I1129 05:32:00.928701 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84e51c8615ea994610df768cbbb5ecbb2835020efb9a2bca864e1455ebfb7b2c"} err="failed to get container status \"84e51c8615ea994610df768cbbb5ecbb2835020efb9a2bca864e1455ebfb7b2c\": rpc error: code = NotFound desc = could not find container \"84e51c8615ea994610df768cbbb5ecbb2835020efb9a2bca864e1455ebfb7b2c\": container with ID starting with 84e51c8615ea994610df768cbbb5ecbb2835020efb9a2bca864e1455ebfb7b2c not found: ID does not exist" Nov 29 05:32:00 crc kubenswrapper[4799]: I1129 05:32:00.928729 4799 scope.go:117] "RemoveContainer" containerID="b815039f205ccbefdb1c4b6dff6fb4970bd0f795ac31ee73beb034cea396124a" Nov 29 05:32:00 crc kubenswrapper[4799]: E1129 05:32:00.929089 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b815039f205ccbefdb1c4b6dff6fb4970bd0f795ac31ee73beb034cea396124a\": container with ID starting with b815039f205ccbefdb1c4b6dff6fb4970bd0f795ac31ee73beb034cea396124a not found: ID does not exist" containerID="b815039f205ccbefdb1c4b6dff6fb4970bd0f795ac31ee73beb034cea396124a" Nov 29 05:32:00 crc kubenswrapper[4799]: I1129 05:32:00.929119 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b815039f205ccbefdb1c4b6dff6fb4970bd0f795ac31ee73beb034cea396124a"} err="failed to get container status \"b815039f205ccbefdb1c4b6dff6fb4970bd0f795ac31ee73beb034cea396124a\": rpc error: code = NotFound desc = could not find container \"b815039f205ccbefdb1c4b6dff6fb4970bd0f795ac31ee73beb034cea396124a\": container with ID starting with b815039f205ccbefdb1c4b6dff6fb4970bd0f795ac31ee73beb034cea396124a not found: ID does not exist" Nov 29 05:32:00 crc kubenswrapper[4799]: I1129 05:32:00.929140 4799 scope.go:117] "RemoveContainer" containerID="6f05c112bd93fd1948fd923d5ac9c763942d7af20ae5056180f11ae141288832" Nov 29 05:32:00 crc kubenswrapper[4799]: E1129 05:32:00.929541 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f05c112bd93fd1948fd923d5ac9c763942d7af20ae5056180f11ae141288832\": container with ID starting with 6f05c112bd93fd1948fd923d5ac9c763942d7af20ae5056180f11ae141288832 not found: ID does not exist" containerID="6f05c112bd93fd1948fd923d5ac9c763942d7af20ae5056180f11ae141288832" Nov 29 05:32:00 crc kubenswrapper[4799]: I1129 05:32:00.929591 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f05c112bd93fd1948fd923d5ac9c763942d7af20ae5056180f11ae141288832"} err="failed to get container status \"6f05c112bd93fd1948fd923d5ac9c763942d7af20ae5056180f11ae141288832\": rpc error: code = NotFound desc = could not find container \"6f05c112bd93fd1948fd923d5ac9c763942d7af20ae5056180f11ae141288832\": container with ID starting with 6f05c112bd93fd1948fd923d5ac9c763942d7af20ae5056180f11ae141288832 not found: ID does not exist" Nov 29 05:32:02 crc kubenswrapper[4799]: I1129 05:32:02.180735 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Nov 29 05:32:02 crc kubenswrapper[4799]: I1129 05:32:02.678030 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="520ff9ee-edbe-4684-925a-c220002c5cc4" path="/var/lib/kubelet/pods/520ff9ee-edbe-4684-925a-c220002c5cc4/volumes" Nov 29 05:32:02 crc kubenswrapper[4799]: I1129 05:32:02.931576 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zf798"] Nov 29 05:32:03 crc kubenswrapper[4799]: I1129 05:32:03.843820 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"c4a30984-7c0c-493b-b752-b603232c837f","Type":"ContainerStarted","Data":"b20a6bbe7fd4332d39c196b6f75d048bdfeaf963c54b3b3499cc6f8dfeb910d9"} Nov 29 05:32:03 crc kubenswrapper[4799]: I1129 05:32:03.844160 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zf798" podUID="723bd6ac-25ca-405b-8418-4e60fa6e2d22" containerName="registry-server" containerID="cri-o://aa6afffe3039808fe86eaf10efc4620f3db866998c3f032070427e362de2cc3b" gracePeriod=2 Nov 29 05:32:03 crc kubenswrapper[4799]: I1129 05:32:03.879975 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.365637928 podStartE2EDuration="54.879944496s" podCreationTimestamp="2025-11-29 05:31:09 +0000 UTC" firstStartedPulling="2025-11-29 05:31:11.662368843 +0000 UTC m=+3147.305299243" lastFinishedPulling="2025-11-29 05:32:02.176675371 +0000 UTC m=+3197.819605811" observedRunningTime="2025-11-29 05:32:03.875041959 +0000 UTC m=+3199.517972379" watchObservedRunningTime="2025-11-29 05:32:03.879944496 +0000 UTC m=+3199.522874906" Nov 29 05:32:04 crc kubenswrapper[4799]: I1129 05:32:04.279007 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zf798" Nov 29 05:32:04 crc kubenswrapper[4799]: I1129 05:32:04.428877 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-849st\" (UniqueName: \"kubernetes.io/projected/723bd6ac-25ca-405b-8418-4e60fa6e2d22-kube-api-access-849st\") pod \"723bd6ac-25ca-405b-8418-4e60fa6e2d22\" (UID: \"723bd6ac-25ca-405b-8418-4e60fa6e2d22\") " Nov 29 05:32:04 crc kubenswrapper[4799]: I1129 05:32:04.429128 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/723bd6ac-25ca-405b-8418-4e60fa6e2d22-catalog-content\") pod \"723bd6ac-25ca-405b-8418-4e60fa6e2d22\" (UID: \"723bd6ac-25ca-405b-8418-4e60fa6e2d22\") " Nov 29 05:32:04 crc kubenswrapper[4799]: I1129 05:32:04.429237 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/723bd6ac-25ca-405b-8418-4e60fa6e2d22-utilities\") pod \"723bd6ac-25ca-405b-8418-4e60fa6e2d22\" (UID: \"723bd6ac-25ca-405b-8418-4e60fa6e2d22\") " Nov 29 05:32:04 crc kubenswrapper[4799]: I1129 05:32:04.430279 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/723bd6ac-25ca-405b-8418-4e60fa6e2d22-utilities" (OuterVolumeSpecName: "utilities") pod "723bd6ac-25ca-405b-8418-4e60fa6e2d22" (UID: "723bd6ac-25ca-405b-8418-4e60fa6e2d22"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:32:04 crc kubenswrapper[4799]: I1129 05:32:04.435329 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/723bd6ac-25ca-405b-8418-4e60fa6e2d22-kube-api-access-849st" (OuterVolumeSpecName: "kube-api-access-849st") pod "723bd6ac-25ca-405b-8418-4e60fa6e2d22" (UID: "723bd6ac-25ca-405b-8418-4e60fa6e2d22"). InnerVolumeSpecName "kube-api-access-849st". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:32:04 crc kubenswrapper[4799]: I1129 05:32:04.488575 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/723bd6ac-25ca-405b-8418-4e60fa6e2d22-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "723bd6ac-25ca-405b-8418-4e60fa6e2d22" (UID: "723bd6ac-25ca-405b-8418-4e60fa6e2d22"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:32:04 crc kubenswrapper[4799]: I1129 05:32:04.532271 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-849st\" (UniqueName: \"kubernetes.io/projected/723bd6ac-25ca-405b-8418-4e60fa6e2d22-kube-api-access-849st\") on node \"crc\" DevicePath \"\"" Nov 29 05:32:04 crc kubenswrapper[4799]: I1129 05:32:04.532334 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/723bd6ac-25ca-405b-8418-4e60fa6e2d22-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 05:32:04 crc kubenswrapper[4799]: I1129 05:32:04.532352 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/723bd6ac-25ca-405b-8418-4e60fa6e2d22-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 05:32:04 crc kubenswrapper[4799]: I1129 05:32:04.857474 4799 generic.go:334] "Generic (PLEG): container finished" podID="723bd6ac-25ca-405b-8418-4e60fa6e2d22" containerID="aa6afffe3039808fe86eaf10efc4620f3db866998c3f032070427e362de2cc3b" exitCode=0 Nov 29 05:32:04 crc kubenswrapper[4799]: I1129 05:32:04.857523 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zf798" event={"ID":"723bd6ac-25ca-405b-8418-4e60fa6e2d22","Type":"ContainerDied","Data":"aa6afffe3039808fe86eaf10efc4620f3db866998c3f032070427e362de2cc3b"} Nov 29 05:32:04 crc kubenswrapper[4799]: I1129 05:32:04.857555 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zf798" event={"ID":"723bd6ac-25ca-405b-8418-4e60fa6e2d22","Type":"ContainerDied","Data":"dcdddb044f12b079026688e77e4269b1b77fc79c6a6a383b28d23c65ee1fb7a2"} Nov 29 05:32:04 crc kubenswrapper[4799]: I1129 05:32:04.857573 4799 scope.go:117] "RemoveContainer" containerID="aa6afffe3039808fe86eaf10efc4620f3db866998c3f032070427e362de2cc3b" Nov 29 05:32:04 crc kubenswrapper[4799]: I1129 05:32:04.857572 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zf798" Nov 29 05:32:04 crc kubenswrapper[4799]: I1129 05:32:04.913325 4799 scope.go:117] "RemoveContainer" containerID="47737218625d4b17dfa86e1e0e19f541fc2c97683e95b025f1667bf14de2d108" Nov 29 05:32:04 crc kubenswrapper[4799]: I1129 05:32:04.923404 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zf798"] Nov 29 05:32:04 crc kubenswrapper[4799]: I1129 05:32:04.941585 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zf798"] Nov 29 05:32:04 crc kubenswrapper[4799]: I1129 05:32:04.946360 4799 scope.go:117] "RemoveContainer" containerID="f64548328f3fa19d3b235dcc7f5bbda8dc20742f77cac05b3e87a2bd8bb4b9a7" Nov 29 05:32:04 crc kubenswrapper[4799]: I1129 05:32:04.991503 4799 scope.go:117] "RemoveContainer" containerID="aa6afffe3039808fe86eaf10efc4620f3db866998c3f032070427e362de2cc3b" Nov 29 05:32:04 crc kubenswrapper[4799]: E1129 05:32:04.992165 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa6afffe3039808fe86eaf10efc4620f3db866998c3f032070427e362de2cc3b\": container with ID starting with aa6afffe3039808fe86eaf10efc4620f3db866998c3f032070427e362de2cc3b not found: ID does not exist" containerID="aa6afffe3039808fe86eaf10efc4620f3db866998c3f032070427e362de2cc3b" Nov 29 05:32:04 crc kubenswrapper[4799]: I1129 05:32:04.992218 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa6afffe3039808fe86eaf10efc4620f3db866998c3f032070427e362de2cc3b"} err="failed to get container status \"aa6afffe3039808fe86eaf10efc4620f3db866998c3f032070427e362de2cc3b\": rpc error: code = NotFound desc = could not find container \"aa6afffe3039808fe86eaf10efc4620f3db866998c3f032070427e362de2cc3b\": container with ID starting with aa6afffe3039808fe86eaf10efc4620f3db866998c3f032070427e362de2cc3b not found: ID does not exist" Nov 29 05:32:04 crc kubenswrapper[4799]: I1129 05:32:04.992249 4799 scope.go:117] "RemoveContainer" containerID="47737218625d4b17dfa86e1e0e19f541fc2c97683e95b025f1667bf14de2d108" Nov 29 05:32:04 crc kubenswrapper[4799]: E1129 05:32:04.992678 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47737218625d4b17dfa86e1e0e19f541fc2c97683e95b025f1667bf14de2d108\": container with ID starting with 47737218625d4b17dfa86e1e0e19f541fc2c97683e95b025f1667bf14de2d108 not found: ID does not exist" containerID="47737218625d4b17dfa86e1e0e19f541fc2c97683e95b025f1667bf14de2d108" Nov 29 05:32:04 crc kubenswrapper[4799]: I1129 05:32:04.992777 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47737218625d4b17dfa86e1e0e19f541fc2c97683e95b025f1667bf14de2d108"} err="failed to get container status \"47737218625d4b17dfa86e1e0e19f541fc2c97683e95b025f1667bf14de2d108\": rpc error: code = NotFound desc = could not find container \"47737218625d4b17dfa86e1e0e19f541fc2c97683e95b025f1667bf14de2d108\": container with ID starting with 47737218625d4b17dfa86e1e0e19f541fc2c97683e95b025f1667bf14de2d108 not found: ID does not exist" Nov 29 05:32:04 crc kubenswrapper[4799]: I1129 05:32:04.992930 4799 scope.go:117] "RemoveContainer" containerID="f64548328f3fa19d3b235dcc7f5bbda8dc20742f77cac05b3e87a2bd8bb4b9a7" Nov 29 05:32:04 crc kubenswrapper[4799]: E1129 05:32:04.993396 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f64548328f3fa19d3b235dcc7f5bbda8dc20742f77cac05b3e87a2bd8bb4b9a7\": container with ID starting with f64548328f3fa19d3b235dcc7f5bbda8dc20742f77cac05b3e87a2bd8bb4b9a7 not found: ID does not exist" containerID="f64548328f3fa19d3b235dcc7f5bbda8dc20742f77cac05b3e87a2bd8bb4b9a7" Nov 29 05:32:04 crc kubenswrapper[4799]: I1129 05:32:04.993483 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f64548328f3fa19d3b235dcc7f5bbda8dc20742f77cac05b3e87a2bd8bb4b9a7"} err="failed to get container status \"f64548328f3fa19d3b235dcc7f5bbda8dc20742f77cac05b3e87a2bd8bb4b9a7\": rpc error: code = NotFound desc = could not find container \"f64548328f3fa19d3b235dcc7f5bbda8dc20742f77cac05b3e87a2bd8bb4b9a7\": container with ID starting with f64548328f3fa19d3b235dcc7f5bbda8dc20742f77cac05b3e87a2bd8bb4b9a7 not found: ID does not exist" Nov 29 05:32:06 crc kubenswrapper[4799]: I1129 05:32:06.674626 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="723bd6ac-25ca-405b-8418-4e60fa6e2d22" path="/var/lib/kubelet/pods/723bd6ac-25ca-405b-8418-4e60fa6e2d22/volumes" Nov 29 05:32:11 crc kubenswrapper[4799]: I1129 05:32:11.659015 4799 scope.go:117] "RemoveContainer" containerID="e6d4b3d1d5b5e1eeb13508d67de3d3c3aac52e76337b55e53793185c0d001d82" Nov 29 05:32:11 crc kubenswrapper[4799]: E1129 05:32:11.660474 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:32:26 crc kubenswrapper[4799]: I1129 05:32:26.687047 4799 scope.go:117] "RemoveContainer" containerID="e6d4b3d1d5b5e1eeb13508d67de3d3c3aac52e76337b55e53793185c0d001d82" Nov 29 05:32:26 crc kubenswrapper[4799]: E1129 05:32:26.689063 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:32:37 crc kubenswrapper[4799]: I1129 05:32:37.660645 4799 scope.go:117] "RemoveContainer" containerID="e6d4b3d1d5b5e1eeb13508d67de3d3c3aac52e76337b55e53793185c0d001d82" Nov 29 05:32:38 crc kubenswrapper[4799]: I1129 05:32:38.301659 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" event={"ID":"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8","Type":"ContainerStarted","Data":"4fb42eada0f495bbc2075ba5dbd595d45c6c10d4fa01b5defcc91b18d95e033f"} Nov 29 05:32:58 crc kubenswrapper[4799]: I1129 05:32:58.381096 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-cnjrf"] Nov 29 05:32:58 crc kubenswrapper[4799]: E1129 05:32:58.382219 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="723bd6ac-25ca-405b-8418-4e60fa6e2d22" containerName="extract-utilities" Nov 29 05:32:58 crc kubenswrapper[4799]: I1129 05:32:58.382235 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="723bd6ac-25ca-405b-8418-4e60fa6e2d22" containerName="extract-utilities" Nov 29 05:32:58 crc kubenswrapper[4799]: E1129 05:32:58.382257 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="723bd6ac-25ca-405b-8418-4e60fa6e2d22" containerName="registry-server" Nov 29 05:32:58 crc kubenswrapper[4799]: I1129 05:32:58.382263 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="723bd6ac-25ca-405b-8418-4e60fa6e2d22" containerName="registry-server" Nov 29 05:32:58 crc kubenswrapper[4799]: E1129 05:32:58.382274 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="520ff9ee-edbe-4684-925a-c220002c5cc4" containerName="registry-server" Nov 29 05:32:58 crc kubenswrapper[4799]: I1129 05:32:58.382280 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="520ff9ee-edbe-4684-925a-c220002c5cc4" containerName="registry-server" Nov 29 05:32:58 crc kubenswrapper[4799]: E1129 05:32:58.382287 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="520ff9ee-edbe-4684-925a-c220002c5cc4" containerName="extract-utilities" Nov 29 05:32:58 crc kubenswrapper[4799]: I1129 05:32:58.382293 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="520ff9ee-edbe-4684-925a-c220002c5cc4" containerName="extract-utilities" Nov 29 05:32:58 crc kubenswrapper[4799]: E1129 05:32:58.382324 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="723bd6ac-25ca-405b-8418-4e60fa6e2d22" containerName="extract-content" Nov 29 05:32:58 crc kubenswrapper[4799]: I1129 05:32:58.382329 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="723bd6ac-25ca-405b-8418-4e60fa6e2d22" containerName="extract-content" Nov 29 05:32:58 crc kubenswrapper[4799]: E1129 05:32:58.382339 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="520ff9ee-edbe-4684-925a-c220002c5cc4" containerName="extract-content" Nov 29 05:32:58 crc kubenswrapper[4799]: I1129 05:32:58.382345 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="520ff9ee-edbe-4684-925a-c220002c5cc4" containerName="extract-content" Nov 29 05:32:58 crc kubenswrapper[4799]: I1129 05:32:58.382522 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="723bd6ac-25ca-405b-8418-4e60fa6e2d22" containerName="registry-server" Nov 29 05:32:58 crc kubenswrapper[4799]: I1129 05:32:58.382537 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="520ff9ee-edbe-4684-925a-c220002c5cc4" containerName="registry-server" Nov 29 05:32:58 crc kubenswrapper[4799]: I1129 05:32:58.384149 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cnjrf" Nov 29 05:32:58 crc kubenswrapper[4799]: I1129 05:32:58.398340 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cnjrf"] Nov 29 05:32:58 crc kubenswrapper[4799]: I1129 05:32:58.435317 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjbpr\" (UniqueName: \"kubernetes.io/projected/e540dc71-7a4e-4aec-871d-830041415a2e-kube-api-access-xjbpr\") pod \"redhat-operators-cnjrf\" (UID: \"e540dc71-7a4e-4aec-871d-830041415a2e\") " pod="openshift-marketplace/redhat-operators-cnjrf" Nov 29 05:32:58 crc kubenswrapper[4799]: I1129 05:32:58.435852 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e540dc71-7a4e-4aec-871d-830041415a2e-utilities\") pod \"redhat-operators-cnjrf\" (UID: \"e540dc71-7a4e-4aec-871d-830041415a2e\") " pod="openshift-marketplace/redhat-operators-cnjrf" Nov 29 05:32:58 crc kubenswrapper[4799]: I1129 05:32:58.436401 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e540dc71-7a4e-4aec-871d-830041415a2e-catalog-content\") pod \"redhat-operators-cnjrf\" (UID: \"e540dc71-7a4e-4aec-871d-830041415a2e\") " pod="openshift-marketplace/redhat-operators-cnjrf" Nov 29 05:32:58 crc kubenswrapper[4799]: I1129 05:32:58.539412 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjbpr\" (UniqueName: \"kubernetes.io/projected/e540dc71-7a4e-4aec-871d-830041415a2e-kube-api-access-xjbpr\") pod \"redhat-operators-cnjrf\" (UID: \"e540dc71-7a4e-4aec-871d-830041415a2e\") " pod="openshift-marketplace/redhat-operators-cnjrf" Nov 29 05:32:58 crc kubenswrapper[4799]: I1129 05:32:58.539549 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e540dc71-7a4e-4aec-871d-830041415a2e-utilities\") pod \"redhat-operators-cnjrf\" (UID: \"e540dc71-7a4e-4aec-871d-830041415a2e\") " pod="openshift-marketplace/redhat-operators-cnjrf" Nov 29 05:32:58 crc kubenswrapper[4799]: I1129 05:32:58.539621 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e540dc71-7a4e-4aec-871d-830041415a2e-catalog-content\") pod \"redhat-operators-cnjrf\" (UID: \"e540dc71-7a4e-4aec-871d-830041415a2e\") " pod="openshift-marketplace/redhat-operators-cnjrf" Nov 29 05:32:58 crc kubenswrapper[4799]: I1129 05:32:58.540368 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e540dc71-7a4e-4aec-871d-830041415a2e-utilities\") pod \"redhat-operators-cnjrf\" (UID: \"e540dc71-7a4e-4aec-871d-830041415a2e\") " pod="openshift-marketplace/redhat-operators-cnjrf" Nov 29 05:32:58 crc kubenswrapper[4799]: I1129 05:32:58.540416 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e540dc71-7a4e-4aec-871d-830041415a2e-catalog-content\") pod \"redhat-operators-cnjrf\" (UID: \"e540dc71-7a4e-4aec-871d-830041415a2e\") " pod="openshift-marketplace/redhat-operators-cnjrf" Nov 29 05:32:58 crc kubenswrapper[4799]: I1129 05:32:58.566652 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjbpr\" (UniqueName: \"kubernetes.io/projected/e540dc71-7a4e-4aec-871d-830041415a2e-kube-api-access-xjbpr\") pod \"redhat-operators-cnjrf\" (UID: \"e540dc71-7a4e-4aec-871d-830041415a2e\") " pod="openshift-marketplace/redhat-operators-cnjrf" Nov 29 05:32:58 crc kubenswrapper[4799]: I1129 05:32:58.706536 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cnjrf" Nov 29 05:32:59 crc kubenswrapper[4799]: I1129 05:32:59.228294 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cnjrf"] Nov 29 05:32:59 crc kubenswrapper[4799]: I1129 05:32:59.664575 4799 generic.go:334] "Generic (PLEG): container finished" podID="e540dc71-7a4e-4aec-871d-830041415a2e" containerID="271bb94275910887ea6f062db4a73c662d087afce033301f2ab063e70a201855" exitCode=0 Nov 29 05:32:59 crc kubenswrapper[4799]: I1129 05:32:59.664674 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cnjrf" event={"ID":"e540dc71-7a4e-4aec-871d-830041415a2e","Type":"ContainerDied","Data":"271bb94275910887ea6f062db4a73c662d087afce033301f2ab063e70a201855"} Nov 29 05:32:59 crc kubenswrapper[4799]: I1129 05:32:59.664923 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cnjrf" event={"ID":"e540dc71-7a4e-4aec-871d-830041415a2e","Type":"ContainerStarted","Data":"fcf375237c5cb8f8fa68fc97829d23e9a8b1f150d7971bf67829b55145f68f3a"} Nov 29 05:33:01 crc kubenswrapper[4799]: I1129 05:33:01.691136 4799 generic.go:334] "Generic (PLEG): container finished" podID="e540dc71-7a4e-4aec-871d-830041415a2e" containerID="7266b6674c3c83116fb249def02fdc76e63647c2aac46609bb2f18e4d08274a4" exitCode=0 Nov 29 05:33:01 crc kubenswrapper[4799]: I1129 05:33:01.691210 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cnjrf" event={"ID":"e540dc71-7a4e-4aec-871d-830041415a2e","Type":"ContainerDied","Data":"7266b6674c3c83116fb249def02fdc76e63647c2aac46609bb2f18e4d08274a4"} Nov 29 05:33:03 crc kubenswrapper[4799]: I1129 05:33:03.716748 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cnjrf" event={"ID":"e540dc71-7a4e-4aec-871d-830041415a2e","Type":"ContainerStarted","Data":"371347a68a2bab0f29f41ae0aab436e5e4f5cdd9e80ae8b3fbf7113c55af698a"} Nov 29 05:33:03 crc kubenswrapper[4799]: I1129 05:33:03.748118 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-cnjrf" podStartSLOduration=2.994585139 podStartE2EDuration="5.748097014s" podCreationTimestamp="2025-11-29 05:32:58 +0000 UTC" firstStartedPulling="2025-11-29 05:32:59.667076371 +0000 UTC m=+3255.310006771" lastFinishedPulling="2025-11-29 05:33:02.420588236 +0000 UTC m=+3258.063518646" observedRunningTime="2025-11-29 05:33:03.742378738 +0000 UTC m=+3259.385309138" watchObservedRunningTime="2025-11-29 05:33:03.748097014 +0000 UTC m=+3259.391027414" Nov 29 05:33:08 crc kubenswrapper[4799]: I1129 05:33:08.707730 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-cnjrf" Nov 29 05:33:08 crc kubenswrapper[4799]: I1129 05:33:08.708603 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-cnjrf" Nov 29 05:33:09 crc kubenswrapper[4799]: I1129 05:33:09.795089 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-cnjrf" podUID="e540dc71-7a4e-4aec-871d-830041415a2e" containerName="registry-server" probeResult="failure" output=< Nov 29 05:33:09 crc kubenswrapper[4799]: timeout: failed to connect service ":50051" within 1s Nov 29 05:33:09 crc kubenswrapper[4799]: > Nov 29 05:33:18 crc kubenswrapper[4799]: I1129 05:33:18.775247 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-cnjrf" Nov 29 05:33:18 crc kubenswrapper[4799]: I1129 05:33:18.859625 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-cnjrf" Nov 29 05:33:19 crc kubenswrapper[4799]: I1129 05:33:19.023156 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cnjrf"] Nov 29 05:33:19 crc kubenswrapper[4799]: I1129 05:33:19.936240 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-cnjrf" podUID="e540dc71-7a4e-4aec-871d-830041415a2e" containerName="registry-server" containerID="cri-o://371347a68a2bab0f29f41ae0aab436e5e4f5cdd9e80ae8b3fbf7113c55af698a" gracePeriod=2 Nov 29 05:33:20 crc kubenswrapper[4799]: I1129 05:33:20.481938 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cnjrf" Nov 29 05:33:20 crc kubenswrapper[4799]: I1129 05:33:20.518471 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e540dc71-7a4e-4aec-871d-830041415a2e-catalog-content\") pod \"e540dc71-7a4e-4aec-871d-830041415a2e\" (UID: \"e540dc71-7a4e-4aec-871d-830041415a2e\") " Nov 29 05:33:20 crc kubenswrapper[4799]: I1129 05:33:20.518849 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xjbpr\" (UniqueName: \"kubernetes.io/projected/e540dc71-7a4e-4aec-871d-830041415a2e-kube-api-access-xjbpr\") pod \"e540dc71-7a4e-4aec-871d-830041415a2e\" (UID: \"e540dc71-7a4e-4aec-871d-830041415a2e\") " Nov 29 05:33:20 crc kubenswrapper[4799]: I1129 05:33:20.518895 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e540dc71-7a4e-4aec-871d-830041415a2e-utilities\") pod \"e540dc71-7a4e-4aec-871d-830041415a2e\" (UID: \"e540dc71-7a4e-4aec-871d-830041415a2e\") " Nov 29 05:33:20 crc kubenswrapper[4799]: I1129 05:33:20.524741 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e540dc71-7a4e-4aec-871d-830041415a2e-utilities" (OuterVolumeSpecName: "utilities") pod "e540dc71-7a4e-4aec-871d-830041415a2e" (UID: "e540dc71-7a4e-4aec-871d-830041415a2e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:33:20 crc kubenswrapper[4799]: I1129 05:33:20.532733 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e540dc71-7a4e-4aec-871d-830041415a2e-kube-api-access-xjbpr" (OuterVolumeSpecName: "kube-api-access-xjbpr") pod "e540dc71-7a4e-4aec-871d-830041415a2e" (UID: "e540dc71-7a4e-4aec-871d-830041415a2e"). InnerVolumeSpecName "kube-api-access-xjbpr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:33:20 crc kubenswrapper[4799]: I1129 05:33:20.622176 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xjbpr\" (UniqueName: \"kubernetes.io/projected/e540dc71-7a4e-4aec-871d-830041415a2e-kube-api-access-xjbpr\") on node \"crc\" DevicePath \"\"" Nov 29 05:33:20 crc kubenswrapper[4799]: I1129 05:33:20.622211 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e540dc71-7a4e-4aec-871d-830041415a2e-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 05:33:20 crc kubenswrapper[4799]: I1129 05:33:20.657682 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e540dc71-7a4e-4aec-871d-830041415a2e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e540dc71-7a4e-4aec-871d-830041415a2e" (UID: "e540dc71-7a4e-4aec-871d-830041415a2e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:33:20 crc kubenswrapper[4799]: I1129 05:33:20.723335 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e540dc71-7a4e-4aec-871d-830041415a2e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 05:33:20 crc kubenswrapper[4799]: I1129 05:33:20.952337 4799 generic.go:334] "Generic (PLEG): container finished" podID="e540dc71-7a4e-4aec-871d-830041415a2e" containerID="371347a68a2bab0f29f41ae0aab436e5e4f5cdd9e80ae8b3fbf7113c55af698a" exitCode=0 Nov 29 05:33:20 crc kubenswrapper[4799]: I1129 05:33:20.952405 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cnjrf" event={"ID":"e540dc71-7a4e-4aec-871d-830041415a2e","Type":"ContainerDied","Data":"371347a68a2bab0f29f41ae0aab436e5e4f5cdd9e80ae8b3fbf7113c55af698a"} Nov 29 05:33:20 crc kubenswrapper[4799]: I1129 05:33:20.952465 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cnjrf" event={"ID":"e540dc71-7a4e-4aec-871d-830041415a2e","Type":"ContainerDied","Data":"fcf375237c5cb8f8fa68fc97829d23e9a8b1f150d7971bf67829b55145f68f3a"} Nov 29 05:33:20 crc kubenswrapper[4799]: I1129 05:33:20.952465 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cnjrf" Nov 29 05:33:20 crc kubenswrapper[4799]: I1129 05:33:20.952493 4799 scope.go:117] "RemoveContainer" containerID="371347a68a2bab0f29f41ae0aab436e5e4f5cdd9e80ae8b3fbf7113c55af698a" Nov 29 05:33:20 crc kubenswrapper[4799]: I1129 05:33:20.986403 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cnjrf"] Nov 29 05:33:20 crc kubenswrapper[4799]: I1129 05:33:20.990642 4799 scope.go:117] "RemoveContainer" containerID="7266b6674c3c83116fb249def02fdc76e63647c2aac46609bb2f18e4d08274a4" Nov 29 05:33:20 crc kubenswrapper[4799]: I1129 05:33:20.994982 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-cnjrf"] Nov 29 05:33:21 crc kubenswrapper[4799]: I1129 05:33:21.020781 4799 scope.go:117] "RemoveContainer" containerID="271bb94275910887ea6f062db4a73c662d087afce033301f2ab063e70a201855" Nov 29 05:33:21 crc kubenswrapper[4799]: I1129 05:33:21.068648 4799 scope.go:117] "RemoveContainer" containerID="371347a68a2bab0f29f41ae0aab436e5e4f5cdd9e80ae8b3fbf7113c55af698a" Nov 29 05:33:21 crc kubenswrapper[4799]: E1129 05:33:21.069753 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"371347a68a2bab0f29f41ae0aab436e5e4f5cdd9e80ae8b3fbf7113c55af698a\": container with ID starting with 371347a68a2bab0f29f41ae0aab436e5e4f5cdd9e80ae8b3fbf7113c55af698a not found: ID does not exist" containerID="371347a68a2bab0f29f41ae0aab436e5e4f5cdd9e80ae8b3fbf7113c55af698a" Nov 29 05:33:21 crc kubenswrapper[4799]: I1129 05:33:21.069875 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"371347a68a2bab0f29f41ae0aab436e5e4f5cdd9e80ae8b3fbf7113c55af698a"} err="failed to get container status \"371347a68a2bab0f29f41ae0aab436e5e4f5cdd9e80ae8b3fbf7113c55af698a\": rpc error: code = NotFound desc = could not find container \"371347a68a2bab0f29f41ae0aab436e5e4f5cdd9e80ae8b3fbf7113c55af698a\": container with ID starting with 371347a68a2bab0f29f41ae0aab436e5e4f5cdd9e80ae8b3fbf7113c55af698a not found: ID does not exist" Nov 29 05:33:21 crc kubenswrapper[4799]: I1129 05:33:21.069928 4799 scope.go:117] "RemoveContainer" containerID="7266b6674c3c83116fb249def02fdc76e63647c2aac46609bb2f18e4d08274a4" Nov 29 05:33:21 crc kubenswrapper[4799]: E1129 05:33:21.070270 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7266b6674c3c83116fb249def02fdc76e63647c2aac46609bb2f18e4d08274a4\": container with ID starting with 7266b6674c3c83116fb249def02fdc76e63647c2aac46609bb2f18e4d08274a4 not found: ID does not exist" containerID="7266b6674c3c83116fb249def02fdc76e63647c2aac46609bb2f18e4d08274a4" Nov 29 05:33:21 crc kubenswrapper[4799]: I1129 05:33:21.070341 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7266b6674c3c83116fb249def02fdc76e63647c2aac46609bb2f18e4d08274a4"} err="failed to get container status \"7266b6674c3c83116fb249def02fdc76e63647c2aac46609bb2f18e4d08274a4\": rpc error: code = NotFound desc = could not find container \"7266b6674c3c83116fb249def02fdc76e63647c2aac46609bb2f18e4d08274a4\": container with ID starting with 7266b6674c3c83116fb249def02fdc76e63647c2aac46609bb2f18e4d08274a4 not found: ID does not exist" Nov 29 05:33:21 crc kubenswrapper[4799]: I1129 05:33:21.070356 4799 scope.go:117] "RemoveContainer" containerID="271bb94275910887ea6f062db4a73c662d087afce033301f2ab063e70a201855" Nov 29 05:33:21 crc kubenswrapper[4799]: E1129 05:33:21.070612 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"271bb94275910887ea6f062db4a73c662d087afce033301f2ab063e70a201855\": container with ID starting with 271bb94275910887ea6f062db4a73c662d087afce033301f2ab063e70a201855 not found: ID does not exist" containerID="271bb94275910887ea6f062db4a73c662d087afce033301f2ab063e70a201855" Nov 29 05:33:21 crc kubenswrapper[4799]: I1129 05:33:21.070650 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"271bb94275910887ea6f062db4a73c662d087afce033301f2ab063e70a201855"} err="failed to get container status \"271bb94275910887ea6f062db4a73c662d087afce033301f2ab063e70a201855\": rpc error: code = NotFound desc = could not find container \"271bb94275910887ea6f062db4a73c662d087afce033301f2ab063e70a201855\": container with ID starting with 271bb94275910887ea6f062db4a73c662d087afce033301f2ab063e70a201855 not found: ID does not exist" Nov 29 05:33:22 crc kubenswrapper[4799]: I1129 05:33:22.672195 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e540dc71-7a4e-4aec-871d-830041415a2e" path="/var/lib/kubelet/pods/e540dc71-7a4e-4aec-871d-830041415a2e/volumes" Nov 29 05:34:37 crc kubenswrapper[4799]: I1129 05:34:37.497980 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 05:34:37 crc kubenswrapper[4799]: I1129 05:34:37.499992 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 05:35:07 crc kubenswrapper[4799]: I1129 05:35:07.497930 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 05:35:07 crc kubenswrapper[4799]: I1129 05:35:07.498434 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 05:35:37 crc kubenswrapper[4799]: I1129 05:35:37.497490 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 05:35:37 crc kubenswrapper[4799]: I1129 05:35:37.498326 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 05:35:37 crc kubenswrapper[4799]: I1129 05:35:37.498387 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-26t88" Nov 29 05:35:37 crc kubenswrapper[4799]: I1129 05:35:37.499391 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4fb42eada0f495bbc2075ba5dbd595d45c6c10d4fa01b5defcc91b18d95e033f"} pod="openshift-machine-config-operator/machine-config-daemon-26t88" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 29 05:35:37 crc kubenswrapper[4799]: I1129 05:35:37.499442 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" containerID="cri-o://4fb42eada0f495bbc2075ba5dbd595d45c6c10d4fa01b5defcc91b18d95e033f" gracePeriod=600 Nov 29 05:35:38 crc kubenswrapper[4799]: I1129 05:35:38.614576 4799 generic.go:334] "Generic (PLEG): container finished" podID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerID="4fb42eada0f495bbc2075ba5dbd595d45c6c10d4fa01b5defcc91b18d95e033f" exitCode=0 Nov 29 05:35:38 crc kubenswrapper[4799]: I1129 05:35:38.614646 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" event={"ID":"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8","Type":"ContainerDied","Data":"4fb42eada0f495bbc2075ba5dbd595d45c6c10d4fa01b5defcc91b18d95e033f"} Nov 29 05:35:38 crc kubenswrapper[4799]: I1129 05:35:38.615237 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" event={"ID":"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8","Type":"ContainerStarted","Data":"7baa15ad529e95d214463d6c72d4de622af57d5098e2ed42c89d7e00788ed77b"} Nov 29 05:35:38 crc kubenswrapper[4799]: I1129 05:35:38.615322 4799 scope.go:117] "RemoveContainer" containerID="e6d4b3d1d5b5e1eeb13508d67de3d3c3aac52e76337b55e53793185c0d001d82" Nov 29 05:35:58 crc kubenswrapper[4799]: I1129 05:35:58.361671 4799 scope.go:117] "RemoveContainer" containerID="ac8cee2b0d8b73b8054fe43eca42d2ee452f3b2acbb44bd14072b0f031ba8b97" Nov 29 05:35:58 crc kubenswrapper[4799]: I1129 05:35:58.384950 4799 scope.go:117] "RemoveContainer" containerID="c9d7ae91be5b663bac35a8f4d01f3ca4da49ccde7511c11f0ab43286d377f6da" Nov 29 05:36:29 crc kubenswrapper[4799]: I1129 05:36:29.476827 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xrmfx"] Nov 29 05:36:29 crc kubenswrapper[4799]: E1129 05:36:29.478544 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e540dc71-7a4e-4aec-871d-830041415a2e" containerName="extract-utilities" Nov 29 05:36:29 crc kubenswrapper[4799]: I1129 05:36:29.478569 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e540dc71-7a4e-4aec-871d-830041415a2e" containerName="extract-utilities" Nov 29 05:36:29 crc kubenswrapper[4799]: E1129 05:36:29.478590 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e540dc71-7a4e-4aec-871d-830041415a2e" containerName="extract-content" Nov 29 05:36:29 crc kubenswrapper[4799]: I1129 05:36:29.478611 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e540dc71-7a4e-4aec-871d-830041415a2e" containerName="extract-content" Nov 29 05:36:29 crc kubenswrapper[4799]: E1129 05:36:29.478640 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e540dc71-7a4e-4aec-871d-830041415a2e" containerName="registry-server" Nov 29 05:36:29 crc kubenswrapper[4799]: I1129 05:36:29.478651 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e540dc71-7a4e-4aec-871d-830041415a2e" containerName="registry-server" Nov 29 05:36:29 crc kubenswrapper[4799]: I1129 05:36:29.478975 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="e540dc71-7a4e-4aec-871d-830041415a2e" containerName="registry-server" Nov 29 05:36:29 crc kubenswrapper[4799]: I1129 05:36:29.480981 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xrmfx" Nov 29 05:36:29 crc kubenswrapper[4799]: I1129 05:36:29.494809 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xrmfx"] Nov 29 05:36:29 crc kubenswrapper[4799]: I1129 05:36:29.499238 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7wx6\" (UniqueName: \"kubernetes.io/projected/11fa6653-bdd5-44e6-a746-e75b56294fbe-kube-api-access-h7wx6\") pod \"redhat-marketplace-xrmfx\" (UID: \"11fa6653-bdd5-44e6-a746-e75b56294fbe\") " pod="openshift-marketplace/redhat-marketplace-xrmfx" Nov 29 05:36:29 crc kubenswrapper[4799]: I1129 05:36:29.499680 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11fa6653-bdd5-44e6-a746-e75b56294fbe-utilities\") pod \"redhat-marketplace-xrmfx\" (UID: \"11fa6653-bdd5-44e6-a746-e75b56294fbe\") " pod="openshift-marketplace/redhat-marketplace-xrmfx" Nov 29 05:36:29 crc kubenswrapper[4799]: I1129 05:36:29.499975 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11fa6653-bdd5-44e6-a746-e75b56294fbe-catalog-content\") pod \"redhat-marketplace-xrmfx\" (UID: \"11fa6653-bdd5-44e6-a746-e75b56294fbe\") " pod="openshift-marketplace/redhat-marketplace-xrmfx" Nov 29 05:36:29 crc kubenswrapper[4799]: I1129 05:36:29.602882 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11fa6653-bdd5-44e6-a746-e75b56294fbe-catalog-content\") pod \"redhat-marketplace-xrmfx\" (UID: \"11fa6653-bdd5-44e6-a746-e75b56294fbe\") " pod="openshift-marketplace/redhat-marketplace-xrmfx" Nov 29 05:36:29 crc kubenswrapper[4799]: I1129 05:36:29.603016 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7wx6\" (UniqueName: \"kubernetes.io/projected/11fa6653-bdd5-44e6-a746-e75b56294fbe-kube-api-access-h7wx6\") pod \"redhat-marketplace-xrmfx\" (UID: \"11fa6653-bdd5-44e6-a746-e75b56294fbe\") " pod="openshift-marketplace/redhat-marketplace-xrmfx" Nov 29 05:36:29 crc kubenswrapper[4799]: I1129 05:36:29.603174 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11fa6653-bdd5-44e6-a746-e75b56294fbe-utilities\") pod \"redhat-marketplace-xrmfx\" (UID: \"11fa6653-bdd5-44e6-a746-e75b56294fbe\") " pod="openshift-marketplace/redhat-marketplace-xrmfx" Nov 29 05:36:29 crc kubenswrapper[4799]: I1129 05:36:29.603652 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11fa6653-bdd5-44e6-a746-e75b56294fbe-catalog-content\") pod \"redhat-marketplace-xrmfx\" (UID: \"11fa6653-bdd5-44e6-a746-e75b56294fbe\") " pod="openshift-marketplace/redhat-marketplace-xrmfx" Nov 29 05:36:29 crc kubenswrapper[4799]: I1129 05:36:29.603784 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11fa6653-bdd5-44e6-a746-e75b56294fbe-utilities\") pod \"redhat-marketplace-xrmfx\" (UID: \"11fa6653-bdd5-44e6-a746-e75b56294fbe\") " pod="openshift-marketplace/redhat-marketplace-xrmfx" Nov 29 05:36:29 crc kubenswrapper[4799]: I1129 05:36:29.630125 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7wx6\" (UniqueName: \"kubernetes.io/projected/11fa6653-bdd5-44e6-a746-e75b56294fbe-kube-api-access-h7wx6\") pod \"redhat-marketplace-xrmfx\" (UID: \"11fa6653-bdd5-44e6-a746-e75b56294fbe\") " pod="openshift-marketplace/redhat-marketplace-xrmfx" Nov 29 05:36:29 crc kubenswrapper[4799]: I1129 05:36:29.808148 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xrmfx" Nov 29 05:36:30 crc kubenswrapper[4799]: I1129 05:36:30.364003 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xrmfx"] Nov 29 05:36:31 crc kubenswrapper[4799]: I1129 05:36:31.176251 4799 generic.go:334] "Generic (PLEG): container finished" podID="11fa6653-bdd5-44e6-a746-e75b56294fbe" containerID="6a94b2a64081b445b11a2178e054f1f59307a5db7f768c7de1d27edd6b6f728d" exitCode=0 Nov 29 05:36:31 crc kubenswrapper[4799]: I1129 05:36:31.176391 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xrmfx" event={"ID":"11fa6653-bdd5-44e6-a746-e75b56294fbe","Type":"ContainerDied","Data":"6a94b2a64081b445b11a2178e054f1f59307a5db7f768c7de1d27edd6b6f728d"} Nov 29 05:36:31 crc kubenswrapper[4799]: I1129 05:36:31.176615 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xrmfx" event={"ID":"11fa6653-bdd5-44e6-a746-e75b56294fbe","Type":"ContainerStarted","Data":"e49cd92ec4d1bec23f9877647c4e7f6b4193a4a65c1d10431862e437aa61575d"} Nov 29 05:36:31 crc kubenswrapper[4799]: I1129 05:36:31.180542 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 29 05:36:33 crc kubenswrapper[4799]: I1129 05:36:33.210308 4799 generic.go:334] "Generic (PLEG): container finished" podID="11fa6653-bdd5-44e6-a746-e75b56294fbe" containerID="eb6c6ff56826b423bd75e73f7f8c29ea6c6fe8b5b90f0f4ca342dfa5931e06ad" exitCode=0 Nov 29 05:36:33 crc kubenswrapper[4799]: I1129 05:36:33.210439 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xrmfx" event={"ID":"11fa6653-bdd5-44e6-a746-e75b56294fbe","Type":"ContainerDied","Data":"eb6c6ff56826b423bd75e73f7f8c29ea6c6fe8b5b90f0f4ca342dfa5931e06ad"} Nov 29 05:36:34 crc kubenswrapper[4799]: I1129 05:36:34.224381 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xrmfx" event={"ID":"11fa6653-bdd5-44e6-a746-e75b56294fbe","Type":"ContainerStarted","Data":"5be27ff4d8112a3a9d3d335a3854b4ee2d82983ed48e735e8a3a61b298d3aaa5"} Nov 29 05:36:34 crc kubenswrapper[4799]: I1129 05:36:34.261023 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xrmfx" podStartSLOduration=2.670935284 podStartE2EDuration="5.260996265s" podCreationTimestamp="2025-11-29 05:36:29 +0000 UTC" firstStartedPulling="2025-11-29 05:36:31.179982542 +0000 UTC m=+3466.822912982" lastFinishedPulling="2025-11-29 05:36:33.770043563 +0000 UTC m=+3469.412973963" observedRunningTime="2025-11-29 05:36:34.253352238 +0000 UTC m=+3469.896317549" watchObservedRunningTime="2025-11-29 05:36:34.260996265 +0000 UTC m=+3469.903926675" Nov 29 05:36:39 crc kubenswrapper[4799]: I1129 05:36:39.808260 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xrmfx" Nov 29 05:36:39 crc kubenswrapper[4799]: I1129 05:36:39.808864 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xrmfx" Nov 29 05:36:39 crc kubenswrapper[4799]: I1129 05:36:39.875233 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xrmfx" Nov 29 05:36:40 crc kubenswrapper[4799]: I1129 05:36:40.360396 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xrmfx" Nov 29 05:36:40 crc kubenswrapper[4799]: I1129 05:36:40.423452 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xrmfx"] Nov 29 05:36:42 crc kubenswrapper[4799]: I1129 05:36:42.322184 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xrmfx" podUID="11fa6653-bdd5-44e6-a746-e75b56294fbe" containerName="registry-server" containerID="cri-o://5be27ff4d8112a3a9d3d335a3854b4ee2d82983ed48e735e8a3a61b298d3aaa5" gracePeriod=2 Nov 29 05:36:42 crc kubenswrapper[4799]: I1129 05:36:42.920954 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xrmfx" Nov 29 05:36:43 crc kubenswrapper[4799]: I1129 05:36:43.012146 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11fa6653-bdd5-44e6-a746-e75b56294fbe-catalog-content\") pod \"11fa6653-bdd5-44e6-a746-e75b56294fbe\" (UID: \"11fa6653-bdd5-44e6-a746-e75b56294fbe\") " Nov 29 05:36:43 crc kubenswrapper[4799]: I1129 05:36:43.012265 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7wx6\" (UniqueName: \"kubernetes.io/projected/11fa6653-bdd5-44e6-a746-e75b56294fbe-kube-api-access-h7wx6\") pod \"11fa6653-bdd5-44e6-a746-e75b56294fbe\" (UID: \"11fa6653-bdd5-44e6-a746-e75b56294fbe\") " Nov 29 05:36:43 crc kubenswrapper[4799]: I1129 05:36:43.012325 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11fa6653-bdd5-44e6-a746-e75b56294fbe-utilities\") pod \"11fa6653-bdd5-44e6-a746-e75b56294fbe\" (UID: \"11fa6653-bdd5-44e6-a746-e75b56294fbe\") " Nov 29 05:36:43 crc kubenswrapper[4799]: I1129 05:36:43.019868 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11fa6653-bdd5-44e6-a746-e75b56294fbe-utilities" (OuterVolumeSpecName: "utilities") pod "11fa6653-bdd5-44e6-a746-e75b56294fbe" (UID: "11fa6653-bdd5-44e6-a746-e75b56294fbe"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:36:43 crc kubenswrapper[4799]: I1129 05:36:43.040891 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11fa6653-bdd5-44e6-a746-e75b56294fbe-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "11fa6653-bdd5-44e6-a746-e75b56294fbe" (UID: "11fa6653-bdd5-44e6-a746-e75b56294fbe"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:36:43 crc kubenswrapper[4799]: I1129 05:36:43.054084 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11fa6653-bdd5-44e6-a746-e75b56294fbe-kube-api-access-h7wx6" (OuterVolumeSpecName: "kube-api-access-h7wx6") pod "11fa6653-bdd5-44e6-a746-e75b56294fbe" (UID: "11fa6653-bdd5-44e6-a746-e75b56294fbe"). InnerVolumeSpecName "kube-api-access-h7wx6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:36:43 crc kubenswrapper[4799]: I1129 05:36:43.115209 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11fa6653-bdd5-44e6-a746-e75b56294fbe-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 05:36:43 crc kubenswrapper[4799]: I1129 05:36:43.115245 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7wx6\" (UniqueName: \"kubernetes.io/projected/11fa6653-bdd5-44e6-a746-e75b56294fbe-kube-api-access-h7wx6\") on node \"crc\" DevicePath \"\"" Nov 29 05:36:43 crc kubenswrapper[4799]: I1129 05:36:43.115260 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11fa6653-bdd5-44e6-a746-e75b56294fbe-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 05:36:43 crc kubenswrapper[4799]: I1129 05:36:43.335859 4799 generic.go:334] "Generic (PLEG): container finished" podID="11fa6653-bdd5-44e6-a746-e75b56294fbe" containerID="5be27ff4d8112a3a9d3d335a3854b4ee2d82983ed48e735e8a3a61b298d3aaa5" exitCode=0 Nov 29 05:36:43 crc kubenswrapper[4799]: I1129 05:36:43.335913 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xrmfx" event={"ID":"11fa6653-bdd5-44e6-a746-e75b56294fbe","Type":"ContainerDied","Data":"5be27ff4d8112a3a9d3d335a3854b4ee2d82983ed48e735e8a3a61b298d3aaa5"} Nov 29 05:36:43 crc kubenswrapper[4799]: I1129 05:36:43.335948 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xrmfx" event={"ID":"11fa6653-bdd5-44e6-a746-e75b56294fbe","Type":"ContainerDied","Data":"e49cd92ec4d1bec23f9877647c4e7f6b4193a4a65c1d10431862e437aa61575d"} Nov 29 05:36:43 crc kubenswrapper[4799]: I1129 05:36:43.335970 4799 scope.go:117] "RemoveContainer" containerID="5be27ff4d8112a3a9d3d335a3854b4ee2d82983ed48e735e8a3a61b298d3aaa5" Nov 29 05:36:43 crc kubenswrapper[4799]: I1129 05:36:43.336134 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xrmfx" Nov 29 05:36:43 crc kubenswrapper[4799]: I1129 05:36:43.378911 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xrmfx"] Nov 29 05:36:43 crc kubenswrapper[4799]: I1129 05:36:43.386228 4799 scope.go:117] "RemoveContainer" containerID="eb6c6ff56826b423bd75e73f7f8c29ea6c6fe8b5b90f0f4ca342dfa5931e06ad" Nov 29 05:36:43 crc kubenswrapper[4799]: I1129 05:36:43.388877 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xrmfx"] Nov 29 05:36:43 crc kubenswrapper[4799]: I1129 05:36:43.417189 4799 scope.go:117] "RemoveContainer" containerID="6a94b2a64081b445b11a2178e054f1f59307a5db7f768c7de1d27edd6b6f728d" Nov 29 05:36:43 crc kubenswrapper[4799]: I1129 05:36:43.489052 4799 scope.go:117] "RemoveContainer" containerID="5be27ff4d8112a3a9d3d335a3854b4ee2d82983ed48e735e8a3a61b298d3aaa5" Nov 29 05:36:43 crc kubenswrapper[4799]: E1129 05:36:43.489533 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5be27ff4d8112a3a9d3d335a3854b4ee2d82983ed48e735e8a3a61b298d3aaa5\": container with ID starting with 5be27ff4d8112a3a9d3d335a3854b4ee2d82983ed48e735e8a3a61b298d3aaa5 not found: ID does not exist" containerID="5be27ff4d8112a3a9d3d335a3854b4ee2d82983ed48e735e8a3a61b298d3aaa5" Nov 29 05:36:43 crc kubenswrapper[4799]: I1129 05:36:43.489617 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5be27ff4d8112a3a9d3d335a3854b4ee2d82983ed48e735e8a3a61b298d3aaa5"} err="failed to get container status \"5be27ff4d8112a3a9d3d335a3854b4ee2d82983ed48e735e8a3a61b298d3aaa5\": rpc error: code = NotFound desc = could not find container \"5be27ff4d8112a3a9d3d335a3854b4ee2d82983ed48e735e8a3a61b298d3aaa5\": container with ID starting with 5be27ff4d8112a3a9d3d335a3854b4ee2d82983ed48e735e8a3a61b298d3aaa5 not found: ID does not exist" Nov 29 05:36:43 crc kubenswrapper[4799]: I1129 05:36:43.489670 4799 scope.go:117] "RemoveContainer" containerID="eb6c6ff56826b423bd75e73f7f8c29ea6c6fe8b5b90f0f4ca342dfa5931e06ad" Nov 29 05:36:43 crc kubenswrapper[4799]: E1129 05:36:43.490247 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb6c6ff56826b423bd75e73f7f8c29ea6c6fe8b5b90f0f4ca342dfa5931e06ad\": container with ID starting with eb6c6ff56826b423bd75e73f7f8c29ea6c6fe8b5b90f0f4ca342dfa5931e06ad not found: ID does not exist" containerID="eb6c6ff56826b423bd75e73f7f8c29ea6c6fe8b5b90f0f4ca342dfa5931e06ad" Nov 29 05:36:43 crc kubenswrapper[4799]: I1129 05:36:43.490282 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb6c6ff56826b423bd75e73f7f8c29ea6c6fe8b5b90f0f4ca342dfa5931e06ad"} err="failed to get container status \"eb6c6ff56826b423bd75e73f7f8c29ea6c6fe8b5b90f0f4ca342dfa5931e06ad\": rpc error: code = NotFound desc = could not find container \"eb6c6ff56826b423bd75e73f7f8c29ea6c6fe8b5b90f0f4ca342dfa5931e06ad\": container with ID starting with eb6c6ff56826b423bd75e73f7f8c29ea6c6fe8b5b90f0f4ca342dfa5931e06ad not found: ID does not exist" Nov 29 05:36:43 crc kubenswrapper[4799]: I1129 05:36:43.490313 4799 scope.go:117] "RemoveContainer" containerID="6a94b2a64081b445b11a2178e054f1f59307a5db7f768c7de1d27edd6b6f728d" Nov 29 05:36:43 crc kubenswrapper[4799]: E1129 05:36:43.490593 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a94b2a64081b445b11a2178e054f1f59307a5db7f768c7de1d27edd6b6f728d\": container with ID starting with 6a94b2a64081b445b11a2178e054f1f59307a5db7f768c7de1d27edd6b6f728d not found: ID does not exist" containerID="6a94b2a64081b445b11a2178e054f1f59307a5db7f768c7de1d27edd6b6f728d" Nov 29 05:36:43 crc kubenswrapper[4799]: I1129 05:36:43.490617 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a94b2a64081b445b11a2178e054f1f59307a5db7f768c7de1d27edd6b6f728d"} err="failed to get container status \"6a94b2a64081b445b11a2178e054f1f59307a5db7f768c7de1d27edd6b6f728d\": rpc error: code = NotFound desc = could not find container \"6a94b2a64081b445b11a2178e054f1f59307a5db7f768c7de1d27edd6b6f728d\": container with ID starting with 6a94b2a64081b445b11a2178e054f1f59307a5db7f768c7de1d27edd6b6f728d not found: ID does not exist" Nov 29 05:36:44 crc kubenswrapper[4799]: I1129 05:36:44.674411 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11fa6653-bdd5-44e6-a746-e75b56294fbe" path="/var/lib/kubelet/pods/11fa6653-bdd5-44e6-a746-e75b56294fbe/volumes" Nov 29 05:37:37 crc kubenswrapper[4799]: I1129 05:37:37.497649 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 05:37:37 crc kubenswrapper[4799]: I1129 05:37:37.498235 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 05:38:07 crc kubenswrapper[4799]: I1129 05:38:07.498215 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 05:38:07 crc kubenswrapper[4799]: I1129 05:38:07.499044 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 05:38:37 crc kubenswrapper[4799]: I1129 05:38:37.497942 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 05:38:37 crc kubenswrapper[4799]: I1129 05:38:37.498438 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 05:38:37 crc kubenswrapper[4799]: I1129 05:38:37.498491 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-26t88" Nov 29 05:38:37 crc kubenswrapper[4799]: I1129 05:38:37.499431 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7baa15ad529e95d214463d6c72d4de622af57d5098e2ed42c89d7e00788ed77b"} pod="openshift-machine-config-operator/machine-config-daemon-26t88" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 29 05:38:37 crc kubenswrapper[4799]: I1129 05:38:37.499487 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" containerID="cri-o://7baa15ad529e95d214463d6c72d4de622af57d5098e2ed42c89d7e00788ed77b" gracePeriod=600 Nov 29 05:38:37 crc kubenswrapper[4799]: E1129 05:38:37.636557 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:38:38 crc kubenswrapper[4799]: I1129 05:38:38.546427 4799 generic.go:334] "Generic (PLEG): container finished" podID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerID="7baa15ad529e95d214463d6c72d4de622af57d5098e2ed42c89d7e00788ed77b" exitCode=0 Nov 29 05:38:38 crc kubenswrapper[4799]: I1129 05:38:38.546489 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" event={"ID":"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8","Type":"ContainerDied","Data":"7baa15ad529e95d214463d6c72d4de622af57d5098e2ed42c89d7e00788ed77b"} Nov 29 05:38:38 crc kubenswrapper[4799]: I1129 05:38:38.546842 4799 scope.go:117] "RemoveContainer" containerID="4fb42eada0f495bbc2075ba5dbd595d45c6c10d4fa01b5defcc91b18d95e033f" Nov 29 05:38:38 crc kubenswrapper[4799]: I1129 05:38:38.547893 4799 scope.go:117] "RemoveContainer" containerID="7baa15ad529e95d214463d6c72d4de622af57d5098e2ed42c89d7e00788ed77b" Nov 29 05:38:38 crc kubenswrapper[4799]: E1129 05:38:38.550553 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:38:48 crc kubenswrapper[4799]: I1129 05:38:48.063844 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-787d-account-create-update-9qmsv"] Nov 29 05:38:48 crc kubenswrapper[4799]: I1129 05:38:48.084560 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-create-r7r7q"] Nov 29 05:38:48 crc kubenswrapper[4799]: I1129 05:38:48.099198 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-787d-account-create-update-9qmsv"] Nov 29 05:38:48 crc kubenswrapper[4799]: I1129 05:38:48.109660 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-create-r7r7q"] Nov 29 05:38:48 crc kubenswrapper[4799]: I1129 05:38:48.674275 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81148106-ce19-4be8-8afd-0db4518977b0" path="/var/lib/kubelet/pods/81148106-ce19-4be8-8afd-0db4518977b0/volumes" Nov 29 05:38:48 crc kubenswrapper[4799]: I1129 05:38:48.675450 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e368006-30f3-4355-86ca-7ac3ec9cbbe7" path="/var/lib/kubelet/pods/9e368006-30f3-4355-86ca-7ac3ec9cbbe7/volumes" Nov 29 05:38:52 crc kubenswrapper[4799]: I1129 05:38:52.659932 4799 scope.go:117] "RemoveContainer" containerID="7baa15ad529e95d214463d6c72d4de622af57d5098e2ed42c89d7e00788ed77b" Nov 29 05:38:52 crc kubenswrapper[4799]: E1129 05:38:52.662653 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:38:58 crc kubenswrapper[4799]: I1129 05:38:58.548098 4799 scope.go:117] "RemoveContainer" containerID="bea20aa939d744e48585ce7b6a148d714c9d1ff15e8300f3fc8c0120a42fbc9c" Nov 29 05:38:58 crc kubenswrapper[4799]: I1129 05:38:58.586844 4799 scope.go:117] "RemoveContainer" containerID="a1d3cd6cc4df20687edb72bd63e3aaa1e58621a04c9675d0f66617bb551cdcb4" Nov 29 05:39:07 crc kubenswrapper[4799]: I1129 05:39:07.659530 4799 scope.go:117] "RemoveContainer" containerID="7baa15ad529e95d214463d6c72d4de622af57d5098e2ed42c89d7e00788ed77b" Nov 29 05:39:07 crc kubenswrapper[4799]: E1129 05:39:07.660430 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:39:19 crc kubenswrapper[4799]: I1129 05:39:19.044881 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-sync-jlhf6"] Nov 29 05:39:19 crc kubenswrapper[4799]: I1129 05:39:19.058163 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-sync-jlhf6"] Nov 29 05:39:20 crc kubenswrapper[4799]: I1129 05:39:20.675602 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a730664-a7b6-4cbd-b657-d6f8ef3803b8" path="/var/lib/kubelet/pods/4a730664-a7b6-4cbd-b657-d6f8ef3803b8/volumes" Nov 29 05:39:21 crc kubenswrapper[4799]: I1129 05:39:21.659014 4799 scope.go:117] "RemoveContainer" containerID="7baa15ad529e95d214463d6c72d4de622af57d5098e2ed42c89d7e00788ed77b" Nov 29 05:39:21 crc kubenswrapper[4799]: E1129 05:39:21.659980 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:39:33 crc kubenswrapper[4799]: I1129 05:39:33.659863 4799 scope.go:117] "RemoveContainer" containerID="7baa15ad529e95d214463d6c72d4de622af57d5098e2ed42c89d7e00788ed77b" Nov 29 05:39:33 crc kubenswrapper[4799]: E1129 05:39:33.660635 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:39:48 crc kubenswrapper[4799]: I1129 05:39:48.659299 4799 scope.go:117] "RemoveContainer" containerID="7baa15ad529e95d214463d6c72d4de622af57d5098e2ed42c89d7e00788ed77b" Nov 29 05:39:48 crc kubenswrapper[4799]: E1129 05:39:48.660170 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:39:58 crc kubenswrapper[4799]: I1129 05:39:58.710127 4799 scope.go:117] "RemoveContainer" containerID="5e6f69a4304cd3351419612f0eabb5b6a202da0cc0f49f1df47483f3354e83e4" Nov 29 05:40:03 crc kubenswrapper[4799]: I1129 05:40:03.659424 4799 scope.go:117] "RemoveContainer" containerID="7baa15ad529e95d214463d6c72d4de622af57d5098e2ed42c89d7e00788ed77b" Nov 29 05:40:03 crc kubenswrapper[4799]: E1129 05:40:03.660197 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:40:14 crc kubenswrapper[4799]: I1129 05:40:14.667289 4799 scope.go:117] "RemoveContainer" containerID="7baa15ad529e95d214463d6c72d4de622af57d5098e2ed42c89d7e00788ed77b" Nov 29 05:40:14 crc kubenswrapper[4799]: E1129 05:40:14.668243 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:40:27 crc kubenswrapper[4799]: I1129 05:40:27.658726 4799 scope.go:117] "RemoveContainer" containerID="7baa15ad529e95d214463d6c72d4de622af57d5098e2ed42c89d7e00788ed77b" Nov 29 05:40:27 crc kubenswrapper[4799]: E1129 05:40:27.659617 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:40:39 crc kubenswrapper[4799]: I1129 05:40:39.659833 4799 scope.go:117] "RemoveContainer" containerID="7baa15ad529e95d214463d6c72d4de622af57d5098e2ed42c89d7e00788ed77b" Nov 29 05:40:39 crc kubenswrapper[4799]: E1129 05:40:39.660755 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:40:50 crc kubenswrapper[4799]: I1129 05:40:50.658668 4799 scope.go:117] "RemoveContainer" containerID="7baa15ad529e95d214463d6c72d4de622af57d5098e2ed42c89d7e00788ed77b" Nov 29 05:40:50 crc kubenswrapper[4799]: E1129 05:40:50.659385 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:41:05 crc kubenswrapper[4799]: I1129 05:41:05.658724 4799 scope.go:117] "RemoveContainer" containerID="7baa15ad529e95d214463d6c72d4de622af57d5098e2ed42c89d7e00788ed77b" Nov 29 05:41:05 crc kubenswrapper[4799]: E1129 05:41:05.659589 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:41:20 crc kubenswrapper[4799]: I1129 05:41:20.663257 4799 scope.go:117] "RemoveContainer" containerID="7baa15ad529e95d214463d6c72d4de622af57d5098e2ed42c89d7e00788ed77b" Nov 29 05:41:20 crc kubenswrapper[4799]: E1129 05:41:20.664333 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:41:33 crc kubenswrapper[4799]: I1129 05:41:33.658943 4799 scope.go:117] "RemoveContainer" containerID="7baa15ad529e95d214463d6c72d4de622af57d5098e2ed42c89d7e00788ed77b" Nov 29 05:41:33 crc kubenswrapper[4799]: E1129 05:41:33.659881 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:41:44 crc kubenswrapper[4799]: I1129 05:41:44.667709 4799 scope.go:117] "RemoveContainer" containerID="7baa15ad529e95d214463d6c72d4de622af57d5098e2ed42c89d7e00788ed77b" Nov 29 05:41:44 crc kubenswrapper[4799]: E1129 05:41:44.669027 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:41:56 crc kubenswrapper[4799]: I1129 05:41:56.659435 4799 scope.go:117] "RemoveContainer" containerID="7baa15ad529e95d214463d6c72d4de622af57d5098e2ed42c89d7e00788ed77b" Nov 29 05:41:56 crc kubenswrapper[4799]: E1129 05:41:56.660180 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:42:10 crc kubenswrapper[4799]: I1129 05:42:10.659062 4799 scope.go:117] "RemoveContainer" containerID="7baa15ad529e95d214463d6c72d4de622af57d5098e2ed42c89d7e00788ed77b" Nov 29 05:42:10 crc kubenswrapper[4799]: E1129 05:42:10.660167 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:42:23 crc kubenswrapper[4799]: I1129 05:42:23.659583 4799 scope.go:117] "RemoveContainer" containerID="7baa15ad529e95d214463d6c72d4de622af57d5098e2ed42c89d7e00788ed77b" Nov 29 05:42:23 crc kubenswrapper[4799]: E1129 05:42:23.661269 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:42:38 crc kubenswrapper[4799]: I1129 05:42:38.658712 4799 scope.go:117] "RemoveContainer" containerID="7baa15ad529e95d214463d6c72d4de622af57d5098e2ed42c89d7e00788ed77b" Nov 29 05:42:38 crc kubenswrapper[4799]: E1129 05:42:38.659459 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:42:39 crc kubenswrapper[4799]: I1129 05:42:39.152669 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nhgzz"] Nov 29 05:42:39 crc kubenswrapper[4799]: E1129 05:42:39.153373 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11fa6653-bdd5-44e6-a746-e75b56294fbe" containerName="registry-server" Nov 29 05:42:39 crc kubenswrapper[4799]: I1129 05:42:39.153392 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="11fa6653-bdd5-44e6-a746-e75b56294fbe" containerName="registry-server" Nov 29 05:42:39 crc kubenswrapper[4799]: E1129 05:42:39.153413 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11fa6653-bdd5-44e6-a746-e75b56294fbe" containerName="extract-utilities" Nov 29 05:42:39 crc kubenswrapper[4799]: I1129 05:42:39.153421 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="11fa6653-bdd5-44e6-a746-e75b56294fbe" containerName="extract-utilities" Nov 29 05:42:39 crc kubenswrapper[4799]: E1129 05:42:39.153438 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11fa6653-bdd5-44e6-a746-e75b56294fbe" containerName="extract-content" Nov 29 05:42:39 crc kubenswrapper[4799]: I1129 05:42:39.153444 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="11fa6653-bdd5-44e6-a746-e75b56294fbe" containerName="extract-content" Nov 29 05:42:39 crc kubenswrapper[4799]: I1129 05:42:39.153644 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="11fa6653-bdd5-44e6-a746-e75b56294fbe" containerName="registry-server" Nov 29 05:42:39 crc kubenswrapper[4799]: I1129 05:42:39.154942 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nhgzz" Nov 29 05:42:39 crc kubenswrapper[4799]: I1129 05:42:39.171504 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkvzl\" (UniqueName: \"kubernetes.io/projected/b7c6f187-070e-4afa-9fa4-85bef02f1ede-kube-api-access-lkvzl\") pod \"certified-operators-nhgzz\" (UID: \"b7c6f187-070e-4afa-9fa4-85bef02f1ede\") " pod="openshift-marketplace/certified-operators-nhgzz" Nov 29 05:42:39 crc kubenswrapper[4799]: I1129 05:42:39.171578 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7c6f187-070e-4afa-9fa4-85bef02f1ede-catalog-content\") pod \"certified-operators-nhgzz\" (UID: \"b7c6f187-070e-4afa-9fa4-85bef02f1ede\") " pod="openshift-marketplace/certified-operators-nhgzz" Nov 29 05:42:39 crc kubenswrapper[4799]: I1129 05:42:39.171728 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7c6f187-070e-4afa-9fa4-85bef02f1ede-utilities\") pod \"certified-operators-nhgzz\" (UID: \"b7c6f187-070e-4afa-9fa4-85bef02f1ede\") " pod="openshift-marketplace/certified-operators-nhgzz" Nov 29 05:42:39 crc kubenswrapper[4799]: I1129 05:42:39.216949 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nhgzz"] Nov 29 05:42:39 crc kubenswrapper[4799]: I1129 05:42:39.273639 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7c6f187-070e-4afa-9fa4-85bef02f1ede-catalog-content\") pod \"certified-operators-nhgzz\" (UID: \"b7c6f187-070e-4afa-9fa4-85bef02f1ede\") " pod="openshift-marketplace/certified-operators-nhgzz" Nov 29 05:42:39 crc kubenswrapper[4799]: I1129 05:42:39.273850 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7c6f187-070e-4afa-9fa4-85bef02f1ede-utilities\") pod \"certified-operators-nhgzz\" (UID: \"b7c6f187-070e-4afa-9fa4-85bef02f1ede\") " pod="openshift-marketplace/certified-operators-nhgzz" Nov 29 05:42:39 crc kubenswrapper[4799]: I1129 05:42:39.273963 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkvzl\" (UniqueName: \"kubernetes.io/projected/b7c6f187-070e-4afa-9fa4-85bef02f1ede-kube-api-access-lkvzl\") pod \"certified-operators-nhgzz\" (UID: \"b7c6f187-070e-4afa-9fa4-85bef02f1ede\") " pod="openshift-marketplace/certified-operators-nhgzz" Nov 29 05:42:39 crc kubenswrapper[4799]: I1129 05:42:39.274443 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7c6f187-070e-4afa-9fa4-85bef02f1ede-utilities\") pod \"certified-operators-nhgzz\" (UID: \"b7c6f187-070e-4afa-9fa4-85bef02f1ede\") " pod="openshift-marketplace/certified-operators-nhgzz" Nov 29 05:42:39 crc kubenswrapper[4799]: I1129 05:42:39.274614 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7c6f187-070e-4afa-9fa4-85bef02f1ede-catalog-content\") pod \"certified-operators-nhgzz\" (UID: \"b7c6f187-070e-4afa-9fa4-85bef02f1ede\") " pod="openshift-marketplace/certified-operators-nhgzz" Nov 29 05:42:39 crc kubenswrapper[4799]: I1129 05:42:39.293394 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkvzl\" (UniqueName: \"kubernetes.io/projected/b7c6f187-070e-4afa-9fa4-85bef02f1ede-kube-api-access-lkvzl\") pod \"certified-operators-nhgzz\" (UID: \"b7c6f187-070e-4afa-9fa4-85bef02f1ede\") " pod="openshift-marketplace/certified-operators-nhgzz" Nov 29 05:42:39 crc kubenswrapper[4799]: I1129 05:42:39.484260 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nhgzz" Nov 29 05:42:39 crc kubenswrapper[4799]: I1129 05:42:39.998598 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nhgzz"] Nov 29 05:42:40 crc kubenswrapper[4799]: I1129 05:42:40.240535 4799 generic.go:334] "Generic (PLEG): container finished" podID="b7c6f187-070e-4afa-9fa4-85bef02f1ede" containerID="853df7412f52b9fceea6af9ef1fe5230fc56ad2f823b94353b580074904c847e" exitCode=0 Nov 29 05:42:40 crc kubenswrapper[4799]: I1129 05:42:40.240833 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nhgzz" event={"ID":"b7c6f187-070e-4afa-9fa4-85bef02f1ede","Type":"ContainerDied","Data":"853df7412f52b9fceea6af9ef1fe5230fc56ad2f823b94353b580074904c847e"} Nov 29 05:42:40 crc kubenswrapper[4799]: I1129 05:42:40.240865 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nhgzz" event={"ID":"b7c6f187-070e-4afa-9fa4-85bef02f1ede","Type":"ContainerStarted","Data":"db82b7d1bbb46c38fbabf7bafed6133a4a34bfc00b6d2153519a1ce4c58a6b65"} Nov 29 05:42:40 crc kubenswrapper[4799]: I1129 05:42:40.243032 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 29 05:42:41 crc kubenswrapper[4799]: I1129 05:42:41.251118 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nhgzz" event={"ID":"b7c6f187-070e-4afa-9fa4-85bef02f1ede","Type":"ContainerStarted","Data":"e4aff6d2ab6aabec198a6bd5fe36535e620886d5565d4f4b02abcd7c43019eb7"} Nov 29 05:42:42 crc kubenswrapper[4799]: I1129 05:42:42.263510 4799 generic.go:334] "Generic (PLEG): container finished" podID="b7c6f187-070e-4afa-9fa4-85bef02f1ede" containerID="e4aff6d2ab6aabec198a6bd5fe36535e620886d5565d4f4b02abcd7c43019eb7" exitCode=0 Nov 29 05:42:42 crc kubenswrapper[4799]: I1129 05:42:42.263564 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nhgzz" event={"ID":"b7c6f187-070e-4afa-9fa4-85bef02f1ede","Type":"ContainerDied","Data":"e4aff6d2ab6aabec198a6bd5fe36535e620886d5565d4f4b02abcd7c43019eb7"} Nov 29 05:42:43 crc kubenswrapper[4799]: I1129 05:42:43.283088 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nhgzz" event={"ID":"b7c6f187-070e-4afa-9fa4-85bef02f1ede","Type":"ContainerStarted","Data":"34762e5c4bbb6b31051e81316a7ba412c09daa85c630853095e2a3761b1ed2a2"} Nov 29 05:42:43 crc kubenswrapper[4799]: I1129 05:42:43.310582 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nhgzz" podStartSLOduration=1.749718867 podStartE2EDuration="4.310562487s" podCreationTimestamp="2025-11-29 05:42:39 +0000 UTC" firstStartedPulling="2025-11-29 05:42:40.242836897 +0000 UTC m=+3835.885767297" lastFinishedPulling="2025-11-29 05:42:42.803680517 +0000 UTC m=+3838.446610917" observedRunningTime="2025-11-29 05:42:43.307292617 +0000 UTC m=+3838.950223037" watchObservedRunningTime="2025-11-29 05:42:43.310562487 +0000 UTC m=+3838.953492887" Nov 29 05:42:49 crc kubenswrapper[4799]: I1129 05:42:49.484602 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nhgzz" Nov 29 05:42:49 crc kubenswrapper[4799]: I1129 05:42:49.485223 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nhgzz" Nov 29 05:42:49 crc kubenswrapper[4799]: I1129 05:42:49.579692 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nhgzz" Nov 29 05:42:49 crc kubenswrapper[4799]: I1129 05:42:49.659453 4799 scope.go:117] "RemoveContainer" containerID="7baa15ad529e95d214463d6c72d4de622af57d5098e2ed42c89d7e00788ed77b" Nov 29 05:42:49 crc kubenswrapper[4799]: E1129 05:42:49.659719 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:42:50 crc kubenswrapper[4799]: I1129 05:42:50.408126 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nhgzz" Nov 29 05:42:50 crc kubenswrapper[4799]: I1129 05:42:50.461443 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nhgzz"] Nov 29 05:42:52 crc kubenswrapper[4799]: I1129 05:42:52.373312 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nhgzz" podUID="b7c6f187-070e-4afa-9fa4-85bef02f1ede" containerName="registry-server" containerID="cri-o://34762e5c4bbb6b31051e81316a7ba412c09daa85c630853095e2a3761b1ed2a2" gracePeriod=2 Nov 29 05:42:52 crc kubenswrapper[4799]: I1129 05:42:52.919334 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nhgzz" Nov 29 05:42:53 crc kubenswrapper[4799]: I1129 05:42:53.065338 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7c6f187-070e-4afa-9fa4-85bef02f1ede-catalog-content\") pod \"b7c6f187-070e-4afa-9fa4-85bef02f1ede\" (UID: \"b7c6f187-070e-4afa-9fa4-85bef02f1ede\") " Nov 29 05:42:53 crc kubenswrapper[4799]: I1129 05:42:53.065467 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lkvzl\" (UniqueName: \"kubernetes.io/projected/b7c6f187-070e-4afa-9fa4-85bef02f1ede-kube-api-access-lkvzl\") pod \"b7c6f187-070e-4afa-9fa4-85bef02f1ede\" (UID: \"b7c6f187-070e-4afa-9fa4-85bef02f1ede\") " Nov 29 05:42:53 crc kubenswrapper[4799]: I1129 05:42:53.065567 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7c6f187-070e-4afa-9fa4-85bef02f1ede-utilities\") pod \"b7c6f187-070e-4afa-9fa4-85bef02f1ede\" (UID: \"b7c6f187-070e-4afa-9fa4-85bef02f1ede\") " Nov 29 05:42:53 crc kubenswrapper[4799]: I1129 05:42:53.066692 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7c6f187-070e-4afa-9fa4-85bef02f1ede-utilities" (OuterVolumeSpecName: "utilities") pod "b7c6f187-070e-4afa-9fa4-85bef02f1ede" (UID: "b7c6f187-070e-4afa-9fa4-85bef02f1ede"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:42:53 crc kubenswrapper[4799]: I1129 05:42:53.071984 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7c6f187-070e-4afa-9fa4-85bef02f1ede-kube-api-access-lkvzl" (OuterVolumeSpecName: "kube-api-access-lkvzl") pod "b7c6f187-070e-4afa-9fa4-85bef02f1ede" (UID: "b7c6f187-070e-4afa-9fa4-85bef02f1ede"). InnerVolumeSpecName "kube-api-access-lkvzl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:42:53 crc kubenswrapper[4799]: I1129 05:42:53.116087 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7c6f187-070e-4afa-9fa4-85bef02f1ede-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b7c6f187-070e-4afa-9fa4-85bef02f1ede" (UID: "b7c6f187-070e-4afa-9fa4-85bef02f1ede"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:42:53 crc kubenswrapper[4799]: I1129 05:42:53.167348 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7c6f187-070e-4afa-9fa4-85bef02f1ede-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 05:42:53 crc kubenswrapper[4799]: I1129 05:42:53.167383 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7c6f187-070e-4afa-9fa4-85bef02f1ede-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 05:42:53 crc kubenswrapper[4799]: I1129 05:42:53.167393 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lkvzl\" (UniqueName: \"kubernetes.io/projected/b7c6f187-070e-4afa-9fa4-85bef02f1ede-kube-api-access-lkvzl\") on node \"crc\" DevicePath \"\"" Nov 29 05:42:53 crc kubenswrapper[4799]: I1129 05:42:53.385364 4799 generic.go:334] "Generic (PLEG): container finished" podID="b7c6f187-070e-4afa-9fa4-85bef02f1ede" containerID="34762e5c4bbb6b31051e81316a7ba412c09daa85c630853095e2a3761b1ed2a2" exitCode=0 Nov 29 05:42:53 crc kubenswrapper[4799]: I1129 05:42:53.385436 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nhgzz" event={"ID":"b7c6f187-070e-4afa-9fa4-85bef02f1ede","Type":"ContainerDied","Data":"34762e5c4bbb6b31051e81316a7ba412c09daa85c630853095e2a3761b1ed2a2"} Nov 29 05:42:53 crc kubenswrapper[4799]: I1129 05:42:53.385473 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nhgzz" Nov 29 05:42:53 crc kubenswrapper[4799]: I1129 05:42:53.385499 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nhgzz" event={"ID":"b7c6f187-070e-4afa-9fa4-85bef02f1ede","Type":"ContainerDied","Data":"db82b7d1bbb46c38fbabf7bafed6133a4a34bfc00b6d2153519a1ce4c58a6b65"} Nov 29 05:42:53 crc kubenswrapper[4799]: I1129 05:42:53.385533 4799 scope.go:117] "RemoveContainer" containerID="34762e5c4bbb6b31051e81316a7ba412c09daa85c630853095e2a3761b1ed2a2" Nov 29 05:42:53 crc kubenswrapper[4799]: I1129 05:42:53.415064 4799 scope.go:117] "RemoveContainer" containerID="e4aff6d2ab6aabec198a6bd5fe36535e620886d5565d4f4b02abcd7c43019eb7" Nov 29 05:42:53 crc kubenswrapper[4799]: I1129 05:42:53.432737 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nhgzz"] Nov 29 05:42:53 crc kubenswrapper[4799]: I1129 05:42:53.442962 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nhgzz"] Nov 29 05:42:53 crc kubenswrapper[4799]: I1129 05:42:53.453917 4799 scope.go:117] "RemoveContainer" containerID="853df7412f52b9fceea6af9ef1fe5230fc56ad2f823b94353b580074904c847e" Nov 29 05:42:53 crc kubenswrapper[4799]: I1129 05:42:53.517307 4799 scope.go:117] "RemoveContainer" containerID="34762e5c4bbb6b31051e81316a7ba412c09daa85c630853095e2a3761b1ed2a2" Nov 29 05:42:53 crc kubenswrapper[4799]: E1129 05:42:53.517695 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34762e5c4bbb6b31051e81316a7ba412c09daa85c630853095e2a3761b1ed2a2\": container with ID starting with 34762e5c4bbb6b31051e81316a7ba412c09daa85c630853095e2a3761b1ed2a2 not found: ID does not exist" containerID="34762e5c4bbb6b31051e81316a7ba412c09daa85c630853095e2a3761b1ed2a2" Nov 29 05:42:53 crc kubenswrapper[4799]: I1129 05:42:53.517825 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34762e5c4bbb6b31051e81316a7ba412c09daa85c630853095e2a3761b1ed2a2"} err="failed to get container status \"34762e5c4bbb6b31051e81316a7ba412c09daa85c630853095e2a3761b1ed2a2\": rpc error: code = NotFound desc = could not find container \"34762e5c4bbb6b31051e81316a7ba412c09daa85c630853095e2a3761b1ed2a2\": container with ID starting with 34762e5c4bbb6b31051e81316a7ba412c09daa85c630853095e2a3761b1ed2a2 not found: ID does not exist" Nov 29 05:42:53 crc kubenswrapper[4799]: I1129 05:42:53.517908 4799 scope.go:117] "RemoveContainer" containerID="e4aff6d2ab6aabec198a6bd5fe36535e620886d5565d4f4b02abcd7c43019eb7" Nov 29 05:42:53 crc kubenswrapper[4799]: E1129 05:42:53.518269 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4aff6d2ab6aabec198a6bd5fe36535e620886d5565d4f4b02abcd7c43019eb7\": container with ID starting with e4aff6d2ab6aabec198a6bd5fe36535e620886d5565d4f4b02abcd7c43019eb7 not found: ID does not exist" containerID="e4aff6d2ab6aabec198a6bd5fe36535e620886d5565d4f4b02abcd7c43019eb7" Nov 29 05:42:53 crc kubenswrapper[4799]: I1129 05:42:53.518293 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4aff6d2ab6aabec198a6bd5fe36535e620886d5565d4f4b02abcd7c43019eb7"} err="failed to get container status \"e4aff6d2ab6aabec198a6bd5fe36535e620886d5565d4f4b02abcd7c43019eb7\": rpc error: code = NotFound desc = could not find container \"e4aff6d2ab6aabec198a6bd5fe36535e620886d5565d4f4b02abcd7c43019eb7\": container with ID starting with e4aff6d2ab6aabec198a6bd5fe36535e620886d5565d4f4b02abcd7c43019eb7 not found: ID does not exist" Nov 29 05:42:53 crc kubenswrapper[4799]: I1129 05:42:53.518308 4799 scope.go:117] "RemoveContainer" containerID="853df7412f52b9fceea6af9ef1fe5230fc56ad2f823b94353b580074904c847e" Nov 29 05:42:53 crc kubenswrapper[4799]: E1129 05:42:53.518637 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"853df7412f52b9fceea6af9ef1fe5230fc56ad2f823b94353b580074904c847e\": container with ID starting with 853df7412f52b9fceea6af9ef1fe5230fc56ad2f823b94353b580074904c847e not found: ID does not exist" containerID="853df7412f52b9fceea6af9ef1fe5230fc56ad2f823b94353b580074904c847e" Nov 29 05:42:53 crc kubenswrapper[4799]: I1129 05:42:53.518709 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"853df7412f52b9fceea6af9ef1fe5230fc56ad2f823b94353b580074904c847e"} err="failed to get container status \"853df7412f52b9fceea6af9ef1fe5230fc56ad2f823b94353b580074904c847e\": rpc error: code = NotFound desc = could not find container \"853df7412f52b9fceea6af9ef1fe5230fc56ad2f823b94353b580074904c847e\": container with ID starting with 853df7412f52b9fceea6af9ef1fe5230fc56ad2f823b94353b580074904c847e not found: ID does not exist" Nov 29 05:42:54 crc kubenswrapper[4799]: I1129 05:42:54.667895 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7c6f187-070e-4afa-9fa4-85bef02f1ede" path="/var/lib/kubelet/pods/b7c6f187-070e-4afa-9fa4-85bef02f1ede/volumes" Nov 29 05:43:01 crc kubenswrapper[4799]: I1129 05:43:01.659133 4799 scope.go:117] "RemoveContainer" containerID="7baa15ad529e95d214463d6c72d4de622af57d5098e2ed42c89d7e00788ed77b" Nov 29 05:43:01 crc kubenswrapper[4799]: E1129 05:43:01.660009 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:43:14 crc kubenswrapper[4799]: I1129 05:43:14.665061 4799 scope.go:117] "RemoveContainer" containerID="7baa15ad529e95d214463d6c72d4de622af57d5098e2ed42c89d7e00788ed77b" Nov 29 05:43:14 crc kubenswrapper[4799]: E1129 05:43:14.665773 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:43:20 crc kubenswrapper[4799]: I1129 05:43:20.587979 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-52vgn"] Nov 29 05:43:20 crc kubenswrapper[4799]: E1129 05:43:20.589089 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7c6f187-070e-4afa-9fa4-85bef02f1ede" containerName="extract-content" Nov 29 05:43:20 crc kubenswrapper[4799]: I1129 05:43:20.589103 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7c6f187-070e-4afa-9fa4-85bef02f1ede" containerName="extract-content" Nov 29 05:43:20 crc kubenswrapper[4799]: E1129 05:43:20.589146 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7c6f187-070e-4afa-9fa4-85bef02f1ede" containerName="registry-server" Nov 29 05:43:20 crc kubenswrapper[4799]: I1129 05:43:20.589152 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7c6f187-070e-4afa-9fa4-85bef02f1ede" containerName="registry-server" Nov 29 05:43:20 crc kubenswrapper[4799]: E1129 05:43:20.589162 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7c6f187-070e-4afa-9fa4-85bef02f1ede" containerName="extract-utilities" Nov 29 05:43:20 crc kubenswrapper[4799]: I1129 05:43:20.589168 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7c6f187-070e-4afa-9fa4-85bef02f1ede" containerName="extract-utilities" Nov 29 05:43:20 crc kubenswrapper[4799]: I1129 05:43:20.589339 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7c6f187-070e-4afa-9fa4-85bef02f1ede" containerName="registry-server" Nov 29 05:43:20 crc kubenswrapper[4799]: I1129 05:43:20.590624 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-52vgn" Nov 29 05:43:20 crc kubenswrapper[4799]: I1129 05:43:20.605648 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-52vgn"] Nov 29 05:43:20 crc kubenswrapper[4799]: I1129 05:43:20.674454 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfvpr\" (UniqueName: \"kubernetes.io/projected/0faee5b1-e347-44b1-9af2-6ef622e8218d-kube-api-access-lfvpr\") pod \"redhat-operators-52vgn\" (UID: \"0faee5b1-e347-44b1-9af2-6ef622e8218d\") " pod="openshift-marketplace/redhat-operators-52vgn" Nov 29 05:43:20 crc kubenswrapper[4799]: I1129 05:43:20.675162 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0faee5b1-e347-44b1-9af2-6ef622e8218d-catalog-content\") pod \"redhat-operators-52vgn\" (UID: \"0faee5b1-e347-44b1-9af2-6ef622e8218d\") " pod="openshift-marketplace/redhat-operators-52vgn" Nov 29 05:43:20 crc kubenswrapper[4799]: I1129 05:43:20.675373 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0faee5b1-e347-44b1-9af2-6ef622e8218d-utilities\") pod \"redhat-operators-52vgn\" (UID: \"0faee5b1-e347-44b1-9af2-6ef622e8218d\") " pod="openshift-marketplace/redhat-operators-52vgn" Nov 29 05:43:20 crc kubenswrapper[4799]: I1129 05:43:20.777195 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0faee5b1-e347-44b1-9af2-6ef622e8218d-utilities\") pod \"redhat-operators-52vgn\" (UID: \"0faee5b1-e347-44b1-9af2-6ef622e8218d\") " pod="openshift-marketplace/redhat-operators-52vgn" Nov 29 05:43:20 crc kubenswrapper[4799]: I1129 05:43:20.777320 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfvpr\" (UniqueName: \"kubernetes.io/projected/0faee5b1-e347-44b1-9af2-6ef622e8218d-kube-api-access-lfvpr\") pod \"redhat-operators-52vgn\" (UID: \"0faee5b1-e347-44b1-9af2-6ef622e8218d\") " pod="openshift-marketplace/redhat-operators-52vgn" Nov 29 05:43:20 crc kubenswrapper[4799]: I1129 05:43:20.777379 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0faee5b1-e347-44b1-9af2-6ef622e8218d-catalog-content\") pod \"redhat-operators-52vgn\" (UID: \"0faee5b1-e347-44b1-9af2-6ef622e8218d\") " pod="openshift-marketplace/redhat-operators-52vgn" Nov 29 05:43:20 crc kubenswrapper[4799]: I1129 05:43:20.777847 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0faee5b1-e347-44b1-9af2-6ef622e8218d-utilities\") pod \"redhat-operators-52vgn\" (UID: \"0faee5b1-e347-44b1-9af2-6ef622e8218d\") " pod="openshift-marketplace/redhat-operators-52vgn" Nov 29 05:43:20 crc kubenswrapper[4799]: I1129 05:43:20.777903 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0faee5b1-e347-44b1-9af2-6ef622e8218d-catalog-content\") pod \"redhat-operators-52vgn\" (UID: \"0faee5b1-e347-44b1-9af2-6ef622e8218d\") " pod="openshift-marketplace/redhat-operators-52vgn" Nov 29 05:43:20 crc kubenswrapper[4799]: I1129 05:43:20.797613 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfvpr\" (UniqueName: \"kubernetes.io/projected/0faee5b1-e347-44b1-9af2-6ef622e8218d-kube-api-access-lfvpr\") pod \"redhat-operators-52vgn\" (UID: \"0faee5b1-e347-44b1-9af2-6ef622e8218d\") " pod="openshift-marketplace/redhat-operators-52vgn" Nov 29 05:43:20 crc kubenswrapper[4799]: I1129 05:43:20.914900 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-52vgn" Nov 29 05:43:21 crc kubenswrapper[4799]: W1129 05:43:21.371185 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0faee5b1_e347_44b1_9af2_6ef622e8218d.slice/crio-f2de795d6bb7d2d31ae4332bc828b349746348d2495fceb341b2a8d01414026b WatchSource:0}: Error finding container f2de795d6bb7d2d31ae4332bc828b349746348d2495fceb341b2a8d01414026b: Status 404 returned error can't find the container with id f2de795d6bb7d2d31ae4332bc828b349746348d2495fceb341b2a8d01414026b Nov 29 05:43:21 crc kubenswrapper[4799]: I1129 05:43:21.371437 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-52vgn"] Nov 29 05:43:21 crc kubenswrapper[4799]: I1129 05:43:21.641171 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-52vgn" event={"ID":"0faee5b1-e347-44b1-9af2-6ef622e8218d","Type":"ContainerStarted","Data":"77998652b41f4e41bcfdfd45bade0de14931eb03bb6573d7425c9363d6fe6d34"} Nov 29 05:43:21 crc kubenswrapper[4799]: I1129 05:43:21.641223 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-52vgn" event={"ID":"0faee5b1-e347-44b1-9af2-6ef622e8218d","Type":"ContainerStarted","Data":"f2de795d6bb7d2d31ae4332bc828b349746348d2495fceb341b2a8d01414026b"} Nov 29 05:43:22 crc kubenswrapper[4799]: I1129 05:43:22.652547 4799 generic.go:334] "Generic (PLEG): container finished" podID="0faee5b1-e347-44b1-9af2-6ef622e8218d" containerID="77998652b41f4e41bcfdfd45bade0de14931eb03bb6573d7425c9363d6fe6d34" exitCode=0 Nov 29 05:43:22 crc kubenswrapper[4799]: I1129 05:43:22.652671 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-52vgn" event={"ID":"0faee5b1-e347-44b1-9af2-6ef622e8218d","Type":"ContainerDied","Data":"77998652b41f4e41bcfdfd45bade0de14931eb03bb6573d7425c9363d6fe6d34"} Nov 29 05:43:23 crc kubenswrapper[4799]: I1129 05:43:23.663713 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-52vgn" event={"ID":"0faee5b1-e347-44b1-9af2-6ef622e8218d","Type":"ContainerStarted","Data":"53543251e64922e8b62d1baf36939fadcc0e459e75b6fbc7bff4b01262df987f"} Nov 29 05:43:24 crc kubenswrapper[4799]: I1129 05:43:24.675194 4799 generic.go:334] "Generic (PLEG): container finished" podID="0faee5b1-e347-44b1-9af2-6ef622e8218d" containerID="53543251e64922e8b62d1baf36939fadcc0e459e75b6fbc7bff4b01262df987f" exitCode=0 Nov 29 05:43:24 crc kubenswrapper[4799]: I1129 05:43:24.675235 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-52vgn" event={"ID":"0faee5b1-e347-44b1-9af2-6ef622e8218d","Type":"ContainerDied","Data":"53543251e64922e8b62d1baf36939fadcc0e459e75b6fbc7bff4b01262df987f"} Nov 29 05:43:25 crc kubenswrapper[4799]: I1129 05:43:25.685886 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-52vgn" event={"ID":"0faee5b1-e347-44b1-9af2-6ef622e8218d","Type":"ContainerStarted","Data":"1347da23f02f966a958ffec195902cf06b3dbcffab3b6211c4e563777cbfb5cc"} Nov 29 05:43:25 crc kubenswrapper[4799]: I1129 05:43:25.710230 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-52vgn" podStartSLOduration=3.261735336 podStartE2EDuration="5.710212298s" podCreationTimestamp="2025-11-29 05:43:20 +0000 UTC" firstStartedPulling="2025-11-29 05:43:22.656630533 +0000 UTC m=+3878.299560943" lastFinishedPulling="2025-11-29 05:43:25.105107475 +0000 UTC m=+3880.748037905" observedRunningTime="2025-11-29 05:43:25.70705272 +0000 UTC m=+3881.349983120" watchObservedRunningTime="2025-11-29 05:43:25.710212298 +0000 UTC m=+3881.353142688" Nov 29 05:43:27 crc kubenswrapper[4799]: I1129 05:43:27.658992 4799 scope.go:117] "RemoveContainer" containerID="7baa15ad529e95d214463d6c72d4de622af57d5098e2ed42c89d7e00788ed77b" Nov 29 05:43:27 crc kubenswrapper[4799]: E1129 05:43:27.659760 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:43:30 crc kubenswrapper[4799]: I1129 05:43:30.916084 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-52vgn" Nov 29 05:43:30 crc kubenswrapper[4799]: I1129 05:43:30.916463 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-52vgn" Nov 29 05:43:30 crc kubenswrapper[4799]: I1129 05:43:30.966103 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-52vgn" Nov 29 05:43:31 crc kubenswrapper[4799]: I1129 05:43:31.830489 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-52vgn" Nov 29 05:43:31 crc kubenswrapper[4799]: I1129 05:43:31.889953 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-52vgn"] Nov 29 05:43:33 crc kubenswrapper[4799]: I1129 05:43:33.800489 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-52vgn" podUID="0faee5b1-e347-44b1-9af2-6ef622e8218d" containerName="registry-server" containerID="cri-o://1347da23f02f966a958ffec195902cf06b3dbcffab3b6211c4e563777cbfb5cc" gracePeriod=2 Nov 29 05:43:34 crc kubenswrapper[4799]: I1129 05:43:34.490021 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-52vgn" Nov 29 05:43:34 crc kubenswrapper[4799]: I1129 05:43:34.553213 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0faee5b1-e347-44b1-9af2-6ef622e8218d-utilities\") pod \"0faee5b1-e347-44b1-9af2-6ef622e8218d\" (UID: \"0faee5b1-e347-44b1-9af2-6ef622e8218d\") " Nov 29 05:43:34 crc kubenswrapper[4799]: I1129 05:43:34.553346 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lfvpr\" (UniqueName: \"kubernetes.io/projected/0faee5b1-e347-44b1-9af2-6ef622e8218d-kube-api-access-lfvpr\") pod \"0faee5b1-e347-44b1-9af2-6ef622e8218d\" (UID: \"0faee5b1-e347-44b1-9af2-6ef622e8218d\") " Nov 29 05:43:34 crc kubenswrapper[4799]: I1129 05:43:34.553376 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0faee5b1-e347-44b1-9af2-6ef622e8218d-catalog-content\") pod \"0faee5b1-e347-44b1-9af2-6ef622e8218d\" (UID: \"0faee5b1-e347-44b1-9af2-6ef622e8218d\") " Nov 29 05:43:34 crc kubenswrapper[4799]: I1129 05:43:34.557525 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0faee5b1-e347-44b1-9af2-6ef622e8218d-utilities" (OuterVolumeSpecName: "utilities") pod "0faee5b1-e347-44b1-9af2-6ef622e8218d" (UID: "0faee5b1-e347-44b1-9af2-6ef622e8218d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:43:34 crc kubenswrapper[4799]: I1129 05:43:34.571587 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0faee5b1-e347-44b1-9af2-6ef622e8218d-kube-api-access-lfvpr" (OuterVolumeSpecName: "kube-api-access-lfvpr") pod "0faee5b1-e347-44b1-9af2-6ef622e8218d" (UID: "0faee5b1-e347-44b1-9af2-6ef622e8218d"). InnerVolumeSpecName "kube-api-access-lfvpr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:43:34 crc kubenswrapper[4799]: I1129 05:43:34.655705 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0faee5b1-e347-44b1-9af2-6ef622e8218d-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 05:43:34 crc kubenswrapper[4799]: I1129 05:43:34.655744 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lfvpr\" (UniqueName: \"kubernetes.io/projected/0faee5b1-e347-44b1-9af2-6ef622e8218d-kube-api-access-lfvpr\") on node \"crc\" DevicePath \"\"" Nov 29 05:43:34 crc kubenswrapper[4799]: I1129 05:43:34.809328 4799 generic.go:334] "Generic (PLEG): container finished" podID="0faee5b1-e347-44b1-9af2-6ef622e8218d" containerID="1347da23f02f966a958ffec195902cf06b3dbcffab3b6211c4e563777cbfb5cc" exitCode=0 Nov 29 05:43:34 crc kubenswrapper[4799]: I1129 05:43:34.809380 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-52vgn" event={"ID":"0faee5b1-e347-44b1-9af2-6ef622e8218d","Type":"ContainerDied","Data":"1347da23f02f966a958ffec195902cf06b3dbcffab3b6211c4e563777cbfb5cc"} Nov 29 05:43:34 crc kubenswrapper[4799]: I1129 05:43:34.809393 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-52vgn" Nov 29 05:43:34 crc kubenswrapper[4799]: I1129 05:43:34.809411 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-52vgn" event={"ID":"0faee5b1-e347-44b1-9af2-6ef622e8218d","Type":"ContainerDied","Data":"f2de795d6bb7d2d31ae4332bc828b349746348d2495fceb341b2a8d01414026b"} Nov 29 05:43:34 crc kubenswrapper[4799]: I1129 05:43:34.809429 4799 scope.go:117] "RemoveContainer" containerID="1347da23f02f966a958ffec195902cf06b3dbcffab3b6211c4e563777cbfb5cc" Nov 29 05:43:34 crc kubenswrapper[4799]: I1129 05:43:34.834199 4799 scope.go:117] "RemoveContainer" containerID="53543251e64922e8b62d1baf36939fadcc0e459e75b6fbc7bff4b01262df987f" Nov 29 05:43:34 crc kubenswrapper[4799]: I1129 05:43:34.867122 4799 scope.go:117] "RemoveContainer" containerID="77998652b41f4e41bcfdfd45bade0de14931eb03bb6573d7425c9363d6fe6d34" Nov 29 05:43:34 crc kubenswrapper[4799]: I1129 05:43:34.916274 4799 scope.go:117] "RemoveContainer" containerID="1347da23f02f966a958ffec195902cf06b3dbcffab3b6211c4e563777cbfb5cc" Nov 29 05:43:34 crc kubenswrapper[4799]: E1129 05:43:34.916765 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1347da23f02f966a958ffec195902cf06b3dbcffab3b6211c4e563777cbfb5cc\": container with ID starting with 1347da23f02f966a958ffec195902cf06b3dbcffab3b6211c4e563777cbfb5cc not found: ID does not exist" containerID="1347da23f02f966a958ffec195902cf06b3dbcffab3b6211c4e563777cbfb5cc" Nov 29 05:43:34 crc kubenswrapper[4799]: I1129 05:43:34.916849 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1347da23f02f966a958ffec195902cf06b3dbcffab3b6211c4e563777cbfb5cc"} err="failed to get container status \"1347da23f02f966a958ffec195902cf06b3dbcffab3b6211c4e563777cbfb5cc\": rpc error: code = NotFound desc = could not find container \"1347da23f02f966a958ffec195902cf06b3dbcffab3b6211c4e563777cbfb5cc\": container with ID starting with 1347da23f02f966a958ffec195902cf06b3dbcffab3b6211c4e563777cbfb5cc not found: ID does not exist" Nov 29 05:43:34 crc kubenswrapper[4799]: I1129 05:43:34.916879 4799 scope.go:117] "RemoveContainer" containerID="53543251e64922e8b62d1baf36939fadcc0e459e75b6fbc7bff4b01262df987f" Nov 29 05:43:34 crc kubenswrapper[4799]: E1129 05:43:34.917189 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53543251e64922e8b62d1baf36939fadcc0e459e75b6fbc7bff4b01262df987f\": container with ID starting with 53543251e64922e8b62d1baf36939fadcc0e459e75b6fbc7bff4b01262df987f not found: ID does not exist" containerID="53543251e64922e8b62d1baf36939fadcc0e459e75b6fbc7bff4b01262df987f" Nov 29 05:43:34 crc kubenswrapper[4799]: I1129 05:43:34.917243 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53543251e64922e8b62d1baf36939fadcc0e459e75b6fbc7bff4b01262df987f"} err="failed to get container status \"53543251e64922e8b62d1baf36939fadcc0e459e75b6fbc7bff4b01262df987f\": rpc error: code = NotFound desc = could not find container \"53543251e64922e8b62d1baf36939fadcc0e459e75b6fbc7bff4b01262df987f\": container with ID starting with 53543251e64922e8b62d1baf36939fadcc0e459e75b6fbc7bff4b01262df987f not found: ID does not exist" Nov 29 05:43:34 crc kubenswrapper[4799]: I1129 05:43:34.917264 4799 scope.go:117] "RemoveContainer" containerID="77998652b41f4e41bcfdfd45bade0de14931eb03bb6573d7425c9363d6fe6d34" Nov 29 05:43:34 crc kubenswrapper[4799]: E1129 05:43:34.917587 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77998652b41f4e41bcfdfd45bade0de14931eb03bb6573d7425c9363d6fe6d34\": container with ID starting with 77998652b41f4e41bcfdfd45bade0de14931eb03bb6573d7425c9363d6fe6d34 not found: ID does not exist" containerID="77998652b41f4e41bcfdfd45bade0de14931eb03bb6573d7425c9363d6fe6d34" Nov 29 05:43:34 crc kubenswrapper[4799]: I1129 05:43:34.917633 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77998652b41f4e41bcfdfd45bade0de14931eb03bb6573d7425c9363d6fe6d34"} err="failed to get container status \"77998652b41f4e41bcfdfd45bade0de14931eb03bb6573d7425c9363d6fe6d34\": rpc error: code = NotFound desc = could not find container \"77998652b41f4e41bcfdfd45bade0de14931eb03bb6573d7425c9363d6fe6d34\": container with ID starting with 77998652b41f4e41bcfdfd45bade0de14931eb03bb6573d7425c9363d6fe6d34 not found: ID does not exist" Nov 29 05:43:36 crc kubenswrapper[4799]: I1129 05:43:36.006695 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0faee5b1-e347-44b1-9af2-6ef622e8218d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0faee5b1-e347-44b1-9af2-6ef622e8218d" (UID: "0faee5b1-e347-44b1-9af2-6ef622e8218d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:43:36 crc kubenswrapper[4799]: I1129 05:43:36.042250 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-52vgn"] Nov 29 05:43:36 crc kubenswrapper[4799]: I1129 05:43:36.079963 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-52vgn"] Nov 29 05:43:36 crc kubenswrapper[4799]: I1129 05:43:36.086628 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0faee5b1-e347-44b1-9af2-6ef622e8218d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 05:43:36 crc kubenswrapper[4799]: I1129 05:43:36.669003 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0faee5b1-e347-44b1-9af2-6ef622e8218d" path="/var/lib/kubelet/pods/0faee5b1-e347-44b1-9af2-6ef622e8218d/volumes" Nov 29 05:43:40 crc kubenswrapper[4799]: I1129 05:43:40.659080 4799 scope.go:117] "RemoveContainer" containerID="7baa15ad529e95d214463d6c72d4de622af57d5098e2ed42c89d7e00788ed77b" Nov 29 05:43:41 crc kubenswrapper[4799]: I1129 05:43:41.876593 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" event={"ID":"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8","Type":"ContainerStarted","Data":"bb3f916f31b5fc3b337d8f5f5faf98dbfdd362c77538109b1107019feab1d3dd"} Nov 29 05:44:30 crc kubenswrapper[4799]: I1129 05:44:30.340882 4799 generic.go:334] "Generic (PLEG): container finished" podID="c4a30984-7c0c-493b-b752-b603232c837f" containerID="b20a6bbe7fd4332d39c196b6f75d048bdfeaf963c54b3b3499cc6f8dfeb910d9" exitCode=0 Nov 29 05:44:30 crc kubenswrapper[4799]: I1129 05:44:30.340956 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"c4a30984-7c0c-493b-b752-b603232c837f","Type":"ContainerDied","Data":"b20a6bbe7fd4332d39c196b6f75d048bdfeaf963c54b3b3499cc6f8dfeb910d9"} Nov 29 05:44:31 crc kubenswrapper[4799]: I1129 05:44:31.849581 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 29 05:44:32 crc kubenswrapper[4799]: I1129 05:44:32.006623 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/c4a30984-7c0c-493b-b752-b603232c837f-test-operator-ephemeral-workdir\") pod \"c4a30984-7c0c-493b-b752-b603232c837f\" (UID: \"c4a30984-7c0c-493b-b752-b603232c837f\") " Nov 29 05:44:32 crc kubenswrapper[4799]: I1129 05:44:32.006729 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c4a30984-7c0c-493b-b752-b603232c837f-ssh-key\") pod \"c4a30984-7c0c-493b-b752-b603232c837f\" (UID: \"c4a30984-7c0c-493b-b752-b603232c837f\") " Nov 29 05:44:32 crc kubenswrapper[4799]: I1129 05:44:32.007961 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"c4a30984-7c0c-493b-b752-b603232c837f\" (UID: \"c4a30984-7c0c-493b-b752-b603232c837f\") " Nov 29 05:44:32 crc kubenswrapper[4799]: I1129 05:44:32.008124 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/c4a30984-7c0c-493b-b752-b603232c837f-ca-certs\") pod \"c4a30984-7c0c-493b-b752-b603232c837f\" (UID: \"c4a30984-7c0c-493b-b752-b603232c837f\") " Nov 29 05:44:32 crc kubenswrapper[4799]: I1129 05:44:32.008248 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/c4a30984-7c0c-493b-b752-b603232c837f-test-operator-ephemeral-temporary\") pod \"c4a30984-7c0c-493b-b752-b603232c837f\" (UID: \"c4a30984-7c0c-493b-b752-b603232c837f\") " Nov 29 05:44:32 crc kubenswrapper[4799]: I1129 05:44:32.008285 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c4a30984-7c0c-493b-b752-b603232c837f-openstack-config-secret\") pod \"c4a30984-7c0c-493b-b752-b603232c837f\" (UID: \"c4a30984-7c0c-493b-b752-b603232c837f\") " Nov 29 05:44:32 crc kubenswrapper[4799]: I1129 05:44:32.008407 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c4a30984-7c0c-493b-b752-b603232c837f-config-data\") pod \"c4a30984-7c0c-493b-b752-b603232c837f\" (UID: \"c4a30984-7c0c-493b-b752-b603232c837f\") " Nov 29 05:44:32 crc kubenswrapper[4799]: I1129 05:44:32.008443 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c4a30984-7c0c-493b-b752-b603232c837f-openstack-config\") pod \"c4a30984-7c0c-493b-b752-b603232c837f\" (UID: \"c4a30984-7c0c-493b-b752-b603232c837f\") " Nov 29 05:44:32 crc kubenswrapper[4799]: I1129 05:44:32.008509 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2fs2z\" (UniqueName: \"kubernetes.io/projected/c4a30984-7c0c-493b-b752-b603232c837f-kube-api-access-2fs2z\") pod \"c4a30984-7c0c-493b-b752-b603232c837f\" (UID: \"c4a30984-7c0c-493b-b752-b603232c837f\") " Nov 29 05:44:32 crc kubenswrapper[4799]: I1129 05:44:32.009779 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4a30984-7c0c-493b-b752-b603232c837f-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "c4a30984-7c0c-493b-b752-b603232c837f" (UID: "c4a30984-7c0c-493b-b752-b603232c837f"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:44:32 crc kubenswrapper[4799]: I1129 05:44:32.010303 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4a30984-7c0c-493b-b752-b603232c837f-config-data" (OuterVolumeSpecName: "config-data") pod "c4a30984-7c0c-493b-b752-b603232c837f" (UID: "c4a30984-7c0c-493b-b752-b603232c837f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 05:44:32 crc kubenswrapper[4799]: I1129 05:44:32.011140 4799 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/c4a30984-7c0c-493b-b752-b603232c837f-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Nov 29 05:44:32 crc kubenswrapper[4799]: I1129 05:44:32.011180 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c4a30984-7c0c-493b-b752-b603232c837f-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 05:44:32 crc kubenswrapper[4799]: I1129 05:44:32.013998 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4a30984-7c0c-493b-b752-b603232c837f-kube-api-access-2fs2z" (OuterVolumeSpecName: "kube-api-access-2fs2z") pod "c4a30984-7c0c-493b-b752-b603232c837f" (UID: "c4a30984-7c0c-493b-b752-b603232c837f"). InnerVolumeSpecName "kube-api-access-2fs2z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:44:32 crc kubenswrapper[4799]: I1129 05:44:32.015124 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "test-operator-logs") pod "c4a30984-7c0c-493b-b752-b603232c837f" (UID: "c4a30984-7c0c-493b-b752-b603232c837f"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 29 05:44:32 crc kubenswrapper[4799]: I1129 05:44:32.015467 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4a30984-7c0c-493b-b752-b603232c837f-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "c4a30984-7c0c-493b-b752-b603232c837f" (UID: "c4a30984-7c0c-493b-b752-b603232c837f"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:44:32 crc kubenswrapper[4799]: I1129 05:44:32.038509 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4a30984-7c0c-493b-b752-b603232c837f-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "c4a30984-7c0c-493b-b752-b603232c837f" (UID: "c4a30984-7c0c-493b-b752-b603232c837f"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:44:32 crc kubenswrapper[4799]: I1129 05:44:32.044015 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4a30984-7c0c-493b-b752-b603232c837f-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "c4a30984-7c0c-493b-b752-b603232c837f" (UID: "c4a30984-7c0c-493b-b752-b603232c837f"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:44:32 crc kubenswrapper[4799]: I1129 05:44:32.048241 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4a30984-7c0c-493b-b752-b603232c837f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c4a30984-7c0c-493b-b752-b603232c837f" (UID: "c4a30984-7c0c-493b-b752-b603232c837f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:44:32 crc kubenswrapper[4799]: I1129 05:44:32.055555 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4a30984-7c0c-493b-b752-b603232c837f-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "c4a30984-7c0c-493b-b752-b603232c837f" (UID: "c4a30984-7c0c-493b-b752-b603232c837f"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 05:44:32 crc kubenswrapper[4799]: I1129 05:44:32.114381 4799 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c4a30984-7c0c-493b-b752-b603232c837f-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 29 05:44:32 crc kubenswrapper[4799]: I1129 05:44:32.114410 4799 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c4a30984-7c0c-493b-b752-b603232c837f-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 29 05:44:32 crc kubenswrapper[4799]: I1129 05:44:32.114419 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2fs2z\" (UniqueName: \"kubernetes.io/projected/c4a30984-7c0c-493b-b752-b603232c837f-kube-api-access-2fs2z\") on node \"crc\" DevicePath \"\"" Nov 29 05:44:32 crc kubenswrapper[4799]: I1129 05:44:32.114428 4799 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/c4a30984-7c0c-493b-b752-b603232c837f-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Nov 29 05:44:32 crc kubenswrapper[4799]: I1129 05:44:32.114440 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c4a30984-7c0c-493b-b752-b603232c837f-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 29 05:44:32 crc kubenswrapper[4799]: I1129 05:44:32.114472 4799 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Nov 29 05:44:32 crc kubenswrapper[4799]: I1129 05:44:32.114482 4799 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/c4a30984-7c0c-493b-b752-b603232c837f-ca-certs\") on node \"crc\" DevicePath \"\"" Nov 29 05:44:32 crc kubenswrapper[4799]: I1129 05:44:32.137924 4799 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Nov 29 05:44:32 crc kubenswrapper[4799]: I1129 05:44:32.215750 4799 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Nov 29 05:44:32 crc kubenswrapper[4799]: I1129 05:44:32.361695 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"c4a30984-7c0c-493b-b752-b603232c837f","Type":"ContainerDied","Data":"614a0889570a9c038cf129ec0198f707974bf7e8c1fc2a5e432330ea1f334508"} Nov 29 05:44:32 crc kubenswrapper[4799]: I1129 05:44:32.362043 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="614a0889570a9c038cf129ec0198f707974bf7e8c1fc2a5e432330ea1f334508" Nov 29 05:44:32 crc kubenswrapper[4799]: I1129 05:44:32.361870 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 29 05:44:36 crc kubenswrapper[4799]: I1129 05:44:36.417449 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 29 05:44:36 crc kubenswrapper[4799]: E1129 05:44:36.418434 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4a30984-7c0c-493b-b752-b603232c837f" containerName="tempest-tests-tempest-tests-runner" Nov 29 05:44:36 crc kubenswrapper[4799]: I1129 05:44:36.418457 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4a30984-7c0c-493b-b752-b603232c837f" containerName="tempest-tests-tempest-tests-runner" Nov 29 05:44:36 crc kubenswrapper[4799]: E1129 05:44:36.418483 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0faee5b1-e347-44b1-9af2-6ef622e8218d" containerName="extract-content" Nov 29 05:44:36 crc kubenswrapper[4799]: I1129 05:44:36.418496 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="0faee5b1-e347-44b1-9af2-6ef622e8218d" containerName="extract-content" Nov 29 05:44:36 crc kubenswrapper[4799]: E1129 05:44:36.418519 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0faee5b1-e347-44b1-9af2-6ef622e8218d" containerName="extract-utilities" Nov 29 05:44:36 crc kubenswrapper[4799]: I1129 05:44:36.418531 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="0faee5b1-e347-44b1-9af2-6ef622e8218d" containerName="extract-utilities" Nov 29 05:44:36 crc kubenswrapper[4799]: E1129 05:44:36.418583 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0faee5b1-e347-44b1-9af2-6ef622e8218d" containerName="registry-server" Nov 29 05:44:36 crc kubenswrapper[4799]: I1129 05:44:36.418595 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="0faee5b1-e347-44b1-9af2-6ef622e8218d" containerName="registry-server" Nov 29 05:44:36 crc kubenswrapper[4799]: I1129 05:44:36.418992 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="0faee5b1-e347-44b1-9af2-6ef622e8218d" containerName="registry-server" Nov 29 05:44:36 crc kubenswrapper[4799]: I1129 05:44:36.419024 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4a30984-7c0c-493b-b752-b603232c837f" containerName="tempest-tests-tempest-tests-runner" Nov 29 05:44:36 crc kubenswrapper[4799]: I1129 05:44:36.420149 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 29 05:44:36 crc kubenswrapper[4799]: I1129 05:44:36.423572 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-n9z7f" Nov 29 05:44:36 crc kubenswrapper[4799]: I1129 05:44:36.432019 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 29 05:44:36 crc kubenswrapper[4799]: I1129 05:44:36.524605 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9vdz\" (UniqueName: \"kubernetes.io/projected/ec2916e5-c59e-4909-9c3c-da11e14e2413-kube-api-access-v9vdz\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"ec2916e5-c59e-4909-9c3c-da11e14e2413\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 29 05:44:36 crc kubenswrapper[4799]: I1129 05:44:36.524741 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"ec2916e5-c59e-4909-9c3c-da11e14e2413\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 29 05:44:36 crc kubenswrapper[4799]: I1129 05:44:36.626397 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9vdz\" (UniqueName: \"kubernetes.io/projected/ec2916e5-c59e-4909-9c3c-da11e14e2413-kube-api-access-v9vdz\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"ec2916e5-c59e-4909-9c3c-da11e14e2413\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 29 05:44:36 crc kubenswrapper[4799]: I1129 05:44:36.626964 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"ec2916e5-c59e-4909-9c3c-da11e14e2413\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 29 05:44:36 crc kubenswrapper[4799]: I1129 05:44:36.627578 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"ec2916e5-c59e-4909-9c3c-da11e14e2413\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 29 05:44:36 crc kubenswrapper[4799]: I1129 05:44:36.660503 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9vdz\" (UniqueName: \"kubernetes.io/projected/ec2916e5-c59e-4909-9c3c-da11e14e2413-kube-api-access-v9vdz\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"ec2916e5-c59e-4909-9c3c-da11e14e2413\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 29 05:44:36 crc kubenswrapper[4799]: I1129 05:44:36.660950 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"ec2916e5-c59e-4909-9c3c-da11e14e2413\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 29 05:44:36 crc kubenswrapper[4799]: I1129 05:44:36.780950 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 29 05:44:37 crc kubenswrapper[4799]: I1129 05:44:37.297015 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 29 05:44:37 crc kubenswrapper[4799]: I1129 05:44:37.411142 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"ec2916e5-c59e-4909-9c3c-da11e14e2413","Type":"ContainerStarted","Data":"02064673149a3cb2fc85ddbadd27e0e78b9864554a8f2cba69426cf24df55790"} Nov 29 05:44:38 crc kubenswrapper[4799]: I1129 05:44:38.423294 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"ec2916e5-c59e-4909-9c3c-da11e14e2413","Type":"ContainerStarted","Data":"c630bbfe8b2604d8b8ac5bc396b55c39c688204db01e62eb89cedb68d2517f16"} Nov 29 05:44:38 crc kubenswrapper[4799]: I1129 05:44:38.441354 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=1.57610139 podStartE2EDuration="2.44133559s" podCreationTimestamp="2025-11-29 05:44:36 +0000 UTC" firstStartedPulling="2025-11-29 05:44:37.309992347 +0000 UTC m=+3952.952922737" lastFinishedPulling="2025-11-29 05:44:38.175226537 +0000 UTC m=+3953.818156937" observedRunningTime="2025-11-29 05:44:38.439768632 +0000 UTC m=+3954.082699032" watchObservedRunningTime="2025-11-29 05:44:38.44133559 +0000 UTC m=+3954.084265990" Nov 29 05:45:00 crc kubenswrapper[4799]: I1129 05:45:00.182950 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406585-b9sbz"] Nov 29 05:45:00 crc kubenswrapper[4799]: I1129 05:45:00.185358 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406585-b9sbz" Nov 29 05:45:00 crc kubenswrapper[4799]: I1129 05:45:00.188489 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 29 05:45:00 crc kubenswrapper[4799]: I1129 05:45:00.188644 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 29 05:45:00 crc kubenswrapper[4799]: I1129 05:45:00.193408 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406585-b9sbz"] Nov 29 05:45:00 crc kubenswrapper[4799]: I1129 05:45:00.340630 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5qwt\" (UniqueName: \"kubernetes.io/projected/a58e5bee-7958-42a4-8f89-f8df4543dc71-kube-api-access-c5qwt\") pod \"collect-profiles-29406585-b9sbz\" (UID: \"a58e5bee-7958-42a4-8f89-f8df4543dc71\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406585-b9sbz" Nov 29 05:45:00 crc kubenswrapper[4799]: I1129 05:45:00.340951 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a58e5bee-7958-42a4-8f89-f8df4543dc71-config-volume\") pod \"collect-profiles-29406585-b9sbz\" (UID: \"a58e5bee-7958-42a4-8f89-f8df4543dc71\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406585-b9sbz" Nov 29 05:45:00 crc kubenswrapper[4799]: I1129 05:45:00.341148 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a58e5bee-7958-42a4-8f89-f8df4543dc71-secret-volume\") pod \"collect-profiles-29406585-b9sbz\" (UID: \"a58e5bee-7958-42a4-8f89-f8df4543dc71\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406585-b9sbz" Nov 29 05:45:00 crc kubenswrapper[4799]: I1129 05:45:00.443005 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5qwt\" (UniqueName: \"kubernetes.io/projected/a58e5bee-7958-42a4-8f89-f8df4543dc71-kube-api-access-c5qwt\") pod \"collect-profiles-29406585-b9sbz\" (UID: \"a58e5bee-7958-42a4-8f89-f8df4543dc71\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406585-b9sbz" Nov 29 05:45:00 crc kubenswrapper[4799]: I1129 05:45:00.443073 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a58e5bee-7958-42a4-8f89-f8df4543dc71-config-volume\") pod \"collect-profiles-29406585-b9sbz\" (UID: \"a58e5bee-7958-42a4-8f89-f8df4543dc71\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406585-b9sbz" Nov 29 05:45:00 crc kubenswrapper[4799]: I1129 05:45:00.443247 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a58e5bee-7958-42a4-8f89-f8df4543dc71-secret-volume\") pod \"collect-profiles-29406585-b9sbz\" (UID: \"a58e5bee-7958-42a4-8f89-f8df4543dc71\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406585-b9sbz" Nov 29 05:45:00 crc kubenswrapper[4799]: I1129 05:45:00.444182 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a58e5bee-7958-42a4-8f89-f8df4543dc71-config-volume\") pod \"collect-profiles-29406585-b9sbz\" (UID: \"a58e5bee-7958-42a4-8f89-f8df4543dc71\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406585-b9sbz" Nov 29 05:45:00 crc kubenswrapper[4799]: I1129 05:45:00.449235 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a58e5bee-7958-42a4-8f89-f8df4543dc71-secret-volume\") pod \"collect-profiles-29406585-b9sbz\" (UID: \"a58e5bee-7958-42a4-8f89-f8df4543dc71\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406585-b9sbz" Nov 29 05:45:00 crc kubenswrapper[4799]: I1129 05:45:00.462911 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5qwt\" (UniqueName: \"kubernetes.io/projected/a58e5bee-7958-42a4-8f89-f8df4543dc71-kube-api-access-c5qwt\") pod \"collect-profiles-29406585-b9sbz\" (UID: \"a58e5bee-7958-42a4-8f89-f8df4543dc71\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406585-b9sbz" Nov 29 05:45:00 crc kubenswrapper[4799]: I1129 05:45:00.509688 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406585-b9sbz" Nov 29 05:45:00 crc kubenswrapper[4799]: I1129 05:45:00.954164 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406585-b9sbz"] Nov 29 05:45:01 crc kubenswrapper[4799]: I1129 05:45:01.649694 4799 generic.go:334] "Generic (PLEG): container finished" podID="a58e5bee-7958-42a4-8f89-f8df4543dc71" containerID="d15eefd2c8649c3bdab50eb4b2c5abcccf07849633c02d7bb90c4348d4ef8638" exitCode=0 Nov 29 05:45:01 crc kubenswrapper[4799]: I1129 05:45:01.649803 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29406585-b9sbz" event={"ID":"a58e5bee-7958-42a4-8f89-f8df4543dc71","Type":"ContainerDied","Data":"d15eefd2c8649c3bdab50eb4b2c5abcccf07849633c02d7bb90c4348d4ef8638"} Nov 29 05:45:01 crc kubenswrapper[4799]: I1129 05:45:01.650091 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29406585-b9sbz" event={"ID":"a58e5bee-7958-42a4-8f89-f8df4543dc71","Type":"ContainerStarted","Data":"f894dec664b35b5cbc716010290ddfd583cd3e943ebe47bb70257c61f4dbe17c"} Nov 29 05:45:03 crc kubenswrapper[4799]: I1129 05:45:03.032430 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406585-b9sbz" Nov 29 05:45:03 crc kubenswrapper[4799]: I1129 05:45:03.106501 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a58e5bee-7958-42a4-8f89-f8df4543dc71-secret-volume\") pod \"a58e5bee-7958-42a4-8f89-f8df4543dc71\" (UID: \"a58e5bee-7958-42a4-8f89-f8df4543dc71\") " Nov 29 05:45:03 crc kubenswrapper[4799]: I1129 05:45:03.107838 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a58e5bee-7958-42a4-8f89-f8df4543dc71-config-volume\") pod \"a58e5bee-7958-42a4-8f89-f8df4543dc71\" (UID: \"a58e5bee-7958-42a4-8f89-f8df4543dc71\") " Nov 29 05:45:03 crc kubenswrapper[4799]: I1129 05:45:03.108085 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c5qwt\" (UniqueName: \"kubernetes.io/projected/a58e5bee-7958-42a4-8f89-f8df4543dc71-kube-api-access-c5qwt\") pod \"a58e5bee-7958-42a4-8f89-f8df4543dc71\" (UID: \"a58e5bee-7958-42a4-8f89-f8df4543dc71\") " Nov 29 05:45:03 crc kubenswrapper[4799]: I1129 05:45:03.108736 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a58e5bee-7958-42a4-8f89-f8df4543dc71-config-volume" (OuterVolumeSpecName: "config-volume") pod "a58e5bee-7958-42a4-8f89-f8df4543dc71" (UID: "a58e5bee-7958-42a4-8f89-f8df4543dc71"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 05:45:03 crc kubenswrapper[4799]: I1129 05:45:03.109518 4799 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a58e5bee-7958-42a4-8f89-f8df4543dc71-config-volume\") on node \"crc\" DevicePath \"\"" Nov 29 05:45:03 crc kubenswrapper[4799]: I1129 05:45:03.112442 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a58e5bee-7958-42a4-8f89-f8df4543dc71-kube-api-access-c5qwt" (OuterVolumeSpecName: "kube-api-access-c5qwt") pod "a58e5bee-7958-42a4-8f89-f8df4543dc71" (UID: "a58e5bee-7958-42a4-8f89-f8df4543dc71"). InnerVolumeSpecName "kube-api-access-c5qwt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:45:03 crc kubenswrapper[4799]: I1129 05:45:03.112436 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a58e5bee-7958-42a4-8f89-f8df4543dc71-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a58e5bee-7958-42a4-8f89-f8df4543dc71" (UID: "a58e5bee-7958-42a4-8f89-f8df4543dc71"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 05:45:03 crc kubenswrapper[4799]: I1129 05:45:03.213289 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c5qwt\" (UniqueName: \"kubernetes.io/projected/a58e5bee-7958-42a4-8f89-f8df4543dc71-kube-api-access-c5qwt\") on node \"crc\" DevicePath \"\"" Nov 29 05:45:03 crc kubenswrapper[4799]: I1129 05:45:03.214444 4799 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a58e5bee-7958-42a4-8f89-f8df4543dc71-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 29 05:45:03 crc kubenswrapper[4799]: I1129 05:45:03.256497 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-ll7z5/must-gather-p24k7"] Nov 29 05:45:03 crc kubenswrapper[4799]: E1129 05:45:03.257146 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a58e5bee-7958-42a4-8f89-f8df4543dc71" containerName="collect-profiles" Nov 29 05:45:03 crc kubenswrapper[4799]: I1129 05:45:03.257166 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a58e5bee-7958-42a4-8f89-f8df4543dc71" containerName="collect-profiles" Nov 29 05:45:03 crc kubenswrapper[4799]: I1129 05:45:03.257406 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="a58e5bee-7958-42a4-8f89-f8df4543dc71" containerName="collect-profiles" Nov 29 05:45:03 crc kubenswrapper[4799]: I1129 05:45:03.258561 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ll7z5/must-gather-p24k7" Nov 29 05:45:03 crc kubenswrapper[4799]: I1129 05:45:03.269685 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-ll7z5/must-gather-p24k7"] Nov 29 05:45:03 crc kubenswrapper[4799]: I1129 05:45:03.280438 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-ll7z5"/"kube-root-ca.crt" Nov 29 05:45:03 crc kubenswrapper[4799]: I1129 05:45:03.280866 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-ll7z5"/"openshift-service-ca.crt" Nov 29 05:45:03 crc kubenswrapper[4799]: I1129 05:45:03.418369 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lt7hr\" (UniqueName: \"kubernetes.io/projected/3fac1ef9-0b5b-49e4-a5df-c95b3982a363-kube-api-access-lt7hr\") pod \"must-gather-p24k7\" (UID: \"3fac1ef9-0b5b-49e4-a5df-c95b3982a363\") " pod="openshift-must-gather-ll7z5/must-gather-p24k7" Nov 29 05:45:03 crc kubenswrapper[4799]: I1129 05:45:03.418494 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3fac1ef9-0b5b-49e4-a5df-c95b3982a363-must-gather-output\") pod \"must-gather-p24k7\" (UID: \"3fac1ef9-0b5b-49e4-a5df-c95b3982a363\") " pod="openshift-must-gather-ll7z5/must-gather-p24k7" Nov 29 05:45:03 crc kubenswrapper[4799]: I1129 05:45:03.520053 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3fac1ef9-0b5b-49e4-a5df-c95b3982a363-must-gather-output\") pod \"must-gather-p24k7\" (UID: \"3fac1ef9-0b5b-49e4-a5df-c95b3982a363\") " pod="openshift-must-gather-ll7z5/must-gather-p24k7" Nov 29 05:45:03 crc kubenswrapper[4799]: I1129 05:45:03.520182 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lt7hr\" (UniqueName: \"kubernetes.io/projected/3fac1ef9-0b5b-49e4-a5df-c95b3982a363-kube-api-access-lt7hr\") pod \"must-gather-p24k7\" (UID: \"3fac1ef9-0b5b-49e4-a5df-c95b3982a363\") " pod="openshift-must-gather-ll7z5/must-gather-p24k7" Nov 29 05:45:03 crc kubenswrapper[4799]: I1129 05:45:03.520636 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3fac1ef9-0b5b-49e4-a5df-c95b3982a363-must-gather-output\") pod \"must-gather-p24k7\" (UID: \"3fac1ef9-0b5b-49e4-a5df-c95b3982a363\") " pod="openshift-must-gather-ll7z5/must-gather-p24k7" Nov 29 05:45:03 crc kubenswrapper[4799]: I1129 05:45:03.544225 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lt7hr\" (UniqueName: \"kubernetes.io/projected/3fac1ef9-0b5b-49e4-a5df-c95b3982a363-kube-api-access-lt7hr\") pod \"must-gather-p24k7\" (UID: \"3fac1ef9-0b5b-49e4-a5df-c95b3982a363\") " pod="openshift-must-gather-ll7z5/must-gather-p24k7" Nov 29 05:45:03 crc kubenswrapper[4799]: I1129 05:45:03.574717 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ll7z5/must-gather-p24k7" Nov 29 05:45:03 crc kubenswrapper[4799]: I1129 05:45:03.679357 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29406585-b9sbz" event={"ID":"a58e5bee-7958-42a4-8f89-f8df4543dc71","Type":"ContainerDied","Data":"f894dec664b35b5cbc716010290ddfd583cd3e943ebe47bb70257c61f4dbe17c"} Nov 29 05:45:03 crc kubenswrapper[4799]: I1129 05:45:03.679601 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f894dec664b35b5cbc716010290ddfd583cd3e943ebe47bb70257c61f4dbe17c" Nov 29 05:45:03 crc kubenswrapper[4799]: I1129 05:45:03.679668 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406585-b9sbz" Nov 29 05:45:04 crc kubenswrapper[4799]: I1129 05:45:04.066172 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-ll7z5/must-gather-p24k7"] Nov 29 05:45:04 crc kubenswrapper[4799]: I1129 05:45:04.126899 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406540-xd5kj"] Nov 29 05:45:04 crc kubenswrapper[4799]: I1129 05:45:04.134562 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406540-xd5kj"] Nov 29 05:45:04 crc kubenswrapper[4799]: I1129 05:45:04.670641 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3ff8087-614c-409e-8238-384a44cdc990" path="/var/lib/kubelet/pods/e3ff8087-614c-409e-8238-384a44cdc990/volumes" Nov 29 05:45:04 crc kubenswrapper[4799]: I1129 05:45:04.688392 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ll7z5/must-gather-p24k7" event={"ID":"3fac1ef9-0b5b-49e4-a5df-c95b3982a363","Type":"ContainerStarted","Data":"0b75ac06ef4559fcf8e0dd5522ab07f830f4baf05657375ebb0ae91294413dc4"} Nov 29 05:45:08 crc kubenswrapper[4799]: I1129 05:45:08.724734 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ll7z5/must-gather-p24k7" event={"ID":"3fac1ef9-0b5b-49e4-a5df-c95b3982a363","Type":"ContainerStarted","Data":"817aa9869e57608dd06023a2defac38f44f2f7ac794773a4e2f4739ed2efc421"} Nov 29 05:45:08 crc kubenswrapper[4799]: I1129 05:45:08.726475 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ll7z5/must-gather-p24k7" event={"ID":"3fac1ef9-0b5b-49e4-a5df-c95b3982a363","Type":"ContainerStarted","Data":"23b13520f503adbd1e25255e588391fa59ac2ebe583d8fd278eebc1a65099261"} Nov 29 05:45:08 crc kubenswrapper[4799]: I1129 05:45:08.750221 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-ll7z5/must-gather-p24k7" podStartSLOduration=1.655914321 podStartE2EDuration="5.750203058s" podCreationTimestamp="2025-11-29 05:45:03 +0000 UTC" firstStartedPulling="2025-11-29 05:45:04.080440998 +0000 UTC m=+3979.723371398" lastFinishedPulling="2025-11-29 05:45:08.174729745 +0000 UTC m=+3983.817660135" observedRunningTime="2025-11-29 05:45:08.743218731 +0000 UTC m=+3984.386149141" watchObservedRunningTime="2025-11-29 05:45:08.750203058 +0000 UTC m=+3984.393133458" Nov 29 05:45:12 crc kubenswrapper[4799]: I1129 05:45:12.275182 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-ll7z5/crc-debug-j2w64"] Nov 29 05:45:12 crc kubenswrapper[4799]: I1129 05:45:12.277284 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ll7z5/crc-debug-j2w64" Nov 29 05:45:12 crc kubenswrapper[4799]: I1129 05:45:12.290907 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-ll7z5"/"default-dockercfg-h574n" Nov 29 05:45:12 crc kubenswrapper[4799]: I1129 05:45:12.415187 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ac2f0929-7fb0-4e7e-b692-e3f1da437287-host\") pod \"crc-debug-j2w64\" (UID: \"ac2f0929-7fb0-4e7e-b692-e3f1da437287\") " pod="openshift-must-gather-ll7z5/crc-debug-j2w64" Nov 29 05:45:12 crc kubenswrapper[4799]: I1129 05:45:12.415231 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxngm\" (UniqueName: \"kubernetes.io/projected/ac2f0929-7fb0-4e7e-b692-e3f1da437287-kube-api-access-wxngm\") pod \"crc-debug-j2w64\" (UID: \"ac2f0929-7fb0-4e7e-b692-e3f1da437287\") " pod="openshift-must-gather-ll7z5/crc-debug-j2w64" Nov 29 05:45:12 crc kubenswrapper[4799]: I1129 05:45:12.517698 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ac2f0929-7fb0-4e7e-b692-e3f1da437287-host\") pod \"crc-debug-j2w64\" (UID: \"ac2f0929-7fb0-4e7e-b692-e3f1da437287\") " pod="openshift-must-gather-ll7z5/crc-debug-j2w64" Nov 29 05:45:12 crc kubenswrapper[4799]: I1129 05:45:12.517737 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxngm\" (UniqueName: \"kubernetes.io/projected/ac2f0929-7fb0-4e7e-b692-e3f1da437287-kube-api-access-wxngm\") pod \"crc-debug-j2w64\" (UID: \"ac2f0929-7fb0-4e7e-b692-e3f1da437287\") " pod="openshift-must-gather-ll7z5/crc-debug-j2w64" Nov 29 05:45:12 crc kubenswrapper[4799]: I1129 05:45:12.517900 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ac2f0929-7fb0-4e7e-b692-e3f1da437287-host\") pod \"crc-debug-j2w64\" (UID: \"ac2f0929-7fb0-4e7e-b692-e3f1da437287\") " pod="openshift-must-gather-ll7z5/crc-debug-j2w64" Nov 29 05:45:12 crc kubenswrapper[4799]: I1129 05:45:12.541851 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxngm\" (UniqueName: \"kubernetes.io/projected/ac2f0929-7fb0-4e7e-b692-e3f1da437287-kube-api-access-wxngm\") pod \"crc-debug-j2w64\" (UID: \"ac2f0929-7fb0-4e7e-b692-e3f1da437287\") " pod="openshift-must-gather-ll7z5/crc-debug-j2w64" Nov 29 05:45:12 crc kubenswrapper[4799]: I1129 05:45:12.598703 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ll7z5/crc-debug-j2w64" Nov 29 05:45:12 crc kubenswrapper[4799]: I1129 05:45:12.759485 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ll7z5/crc-debug-j2w64" event={"ID":"ac2f0929-7fb0-4e7e-b692-e3f1da437287","Type":"ContainerStarted","Data":"60c9c331de3746ba2663f5cd576b394b6b08142d90c7ff7103b7bc1bfbb8cccd"} Nov 29 05:45:23 crc kubenswrapper[4799]: I1129 05:45:23.852440 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ll7z5/crc-debug-j2w64" event={"ID":"ac2f0929-7fb0-4e7e-b692-e3f1da437287","Type":"ContainerStarted","Data":"ce12a605982e76f68a3be954db1c235af8e583a9c7ae7885629e30d988a8a4c3"} Nov 29 05:45:23 crc kubenswrapper[4799]: I1129 05:45:23.884478 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-ll7z5/crc-debug-j2w64" podStartSLOduration=1.447838765 podStartE2EDuration="11.88445488s" podCreationTimestamp="2025-11-29 05:45:12 +0000 UTC" firstStartedPulling="2025-11-29 05:45:12.643562347 +0000 UTC m=+3988.286492747" lastFinishedPulling="2025-11-29 05:45:23.080178462 +0000 UTC m=+3998.723108862" observedRunningTime="2025-11-29 05:45:23.870342213 +0000 UTC m=+3999.513272613" watchObservedRunningTime="2025-11-29 05:45:23.88445488 +0000 UTC m=+3999.527385280" Nov 29 05:45:26 crc kubenswrapper[4799]: I1129 05:45:26.722036 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-wc7hk"] Nov 29 05:45:26 crc kubenswrapper[4799]: I1129 05:45:26.725134 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wc7hk" Nov 29 05:45:26 crc kubenswrapper[4799]: I1129 05:45:26.745860 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wc7hk"] Nov 29 05:45:26 crc kubenswrapper[4799]: I1129 05:45:26.797277 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thbmc\" (UniqueName: \"kubernetes.io/projected/3642ba6f-6029-462d-9441-903056f9886a-kube-api-access-thbmc\") pod \"community-operators-wc7hk\" (UID: \"3642ba6f-6029-462d-9441-903056f9886a\") " pod="openshift-marketplace/community-operators-wc7hk" Nov 29 05:45:26 crc kubenswrapper[4799]: I1129 05:45:26.798125 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3642ba6f-6029-462d-9441-903056f9886a-utilities\") pod \"community-operators-wc7hk\" (UID: \"3642ba6f-6029-462d-9441-903056f9886a\") " pod="openshift-marketplace/community-operators-wc7hk" Nov 29 05:45:26 crc kubenswrapper[4799]: I1129 05:45:26.798392 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3642ba6f-6029-462d-9441-903056f9886a-catalog-content\") pod \"community-operators-wc7hk\" (UID: \"3642ba6f-6029-462d-9441-903056f9886a\") " pod="openshift-marketplace/community-operators-wc7hk" Nov 29 05:45:26 crc kubenswrapper[4799]: I1129 05:45:26.900373 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thbmc\" (UniqueName: \"kubernetes.io/projected/3642ba6f-6029-462d-9441-903056f9886a-kube-api-access-thbmc\") pod \"community-operators-wc7hk\" (UID: \"3642ba6f-6029-462d-9441-903056f9886a\") " pod="openshift-marketplace/community-operators-wc7hk" Nov 29 05:45:26 crc kubenswrapper[4799]: I1129 05:45:26.900433 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3642ba6f-6029-462d-9441-903056f9886a-utilities\") pod \"community-operators-wc7hk\" (UID: \"3642ba6f-6029-462d-9441-903056f9886a\") " pod="openshift-marketplace/community-operators-wc7hk" Nov 29 05:45:26 crc kubenswrapper[4799]: I1129 05:45:26.900471 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3642ba6f-6029-462d-9441-903056f9886a-catalog-content\") pod \"community-operators-wc7hk\" (UID: \"3642ba6f-6029-462d-9441-903056f9886a\") " pod="openshift-marketplace/community-operators-wc7hk" Nov 29 05:45:26 crc kubenswrapper[4799]: I1129 05:45:26.901108 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3642ba6f-6029-462d-9441-903056f9886a-catalog-content\") pod \"community-operators-wc7hk\" (UID: \"3642ba6f-6029-462d-9441-903056f9886a\") " pod="openshift-marketplace/community-operators-wc7hk" Nov 29 05:45:26 crc kubenswrapper[4799]: I1129 05:45:26.901731 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3642ba6f-6029-462d-9441-903056f9886a-utilities\") pod \"community-operators-wc7hk\" (UID: \"3642ba6f-6029-462d-9441-903056f9886a\") " pod="openshift-marketplace/community-operators-wc7hk" Nov 29 05:45:26 crc kubenswrapper[4799]: I1129 05:45:26.923782 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thbmc\" (UniqueName: \"kubernetes.io/projected/3642ba6f-6029-462d-9441-903056f9886a-kube-api-access-thbmc\") pod \"community-operators-wc7hk\" (UID: \"3642ba6f-6029-462d-9441-903056f9886a\") " pod="openshift-marketplace/community-operators-wc7hk" Nov 29 05:45:27 crc kubenswrapper[4799]: I1129 05:45:27.061017 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wc7hk" Nov 29 05:45:27 crc kubenswrapper[4799]: I1129 05:45:27.666477 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wc7hk"] Nov 29 05:45:27 crc kubenswrapper[4799]: W1129 05:45:27.670099 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3642ba6f_6029_462d_9441_903056f9886a.slice/crio-06420f5ed46c3370710ea8de1bc303da55aed6cceb4a07a27f2ecaf4dd8b4f0f WatchSource:0}: Error finding container 06420f5ed46c3370710ea8de1bc303da55aed6cceb4a07a27f2ecaf4dd8b4f0f: Status 404 returned error can't find the container with id 06420f5ed46c3370710ea8de1bc303da55aed6cceb4a07a27f2ecaf4dd8b4f0f Nov 29 05:45:27 crc kubenswrapper[4799]: I1129 05:45:27.908979 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wc7hk" event={"ID":"3642ba6f-6029-462d-9441-903056f9886a","Type":"ContainerStarted","Data":"06420f5ed46c3370710ea8de1bc303da55aed6cceb4a07a27f2ecaf4dd8b4f0f"} Nov 29 05:45:30 crc kubenswrapper[4799]: I1129 05:45:30.938701 4799 generic.go:334] "Generic (PLEG): container finished" podID="3642ba6f-6029-462d-9441-903056f9886a" containerID="93122f0b2a043a21cfcb286809cfdf1e873b30e9a93a0637bd782888122e1605" exitCode=0 Nov 29 05:45:30 crc kubenswrapper[4799]: I1129 05:45:30.939960 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wc7hk" event={"ID":"3642ba6f-6029-462d-9441-903056f9886a","Type":"ContainerDied","Data":"93122f0b2a043a21cfcb286809cfdf1e873b30e9a93a0637bd782888122e1605"} Nov 29 05:45:38 crc kubenswrapper[4799]: I1129 05:45:38.026095 4799 generic.go:334] "Generic (PLEG): container finished" podID="3642ba6f-6029-462d-9441-903056f9886a" containerID="36d6da82df9fcfbbef4e3bc93398b5aec2ac19346fbae5ae17d76a4d7f169706" exitCode=0 Nov 29 05:45:38 crc kubenswrapper[4799]: I1129 05:45:38.026200 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wc7hk" event={"ID":"3642ba6f-6029-462d-9441-903056f9886a","Type":"ContainerDied","Data":"36d6da82df9fcfbbef4e3bc93398b5aec2ac19346fbae5ae17d76a4d7f169706"} Nov 29 05:45:40 crc kubenswrapper[4799]: I1129 05:45:40.053543 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wc7hk" event={"ID":"3642ba6f-6029-462d-9441-903056f9886a","Type":"ContainerStarted","Data":"76fd9f29ecaf3ce233a26c04bd001769defd788749acfe470231d53809ac1711"} Nov 29 05:45:47 crc kubenswrapper[4799]: I1129 05:45:47.063060 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-wc7hk" Nov 29 05:45:47 crc kubenswrapper[4799]: I1129 05:45:47.063613 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-wc7hk" Nov 29 05:45:47 crc kubenswrapper[4799]: I1129 05:45:47.158895 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-wc7hk" Nov 29 05:45:47 crc kubenswrapper[4799]: I1129 05:45:47.199003 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-wc7hk" podStartSLOduration=13.424331601 podStartE2EDuration="21.198954767s" podCreationTimestamp="2025-11-29 05:45:26 +0000 UTC" firstStartedPulling="2025-11-29 05:45:30.942026757 +0000 UTC m=+4006.584957157" lastFinishedPulling="2025-11-29 05:45:38.716649923 +0000 UTC m=+4014.359580323" observedRunningTime="2025-11-29 05:45:40.084672533 +0000 UTC m=+4015.727602953" watchObservedRunningTime="2025-11-29 05:45:47.198954767 +0000 UTC m=+4022.841885167" Nov 29 05:45:47 crc kubenswrapper[4799]: I1129 05:45:47.224296 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-wc7hk" Nov 29 05:45:47 crc kubenswrapper[4799]: I1129 05:45:47.306686 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wc7hk"] Nov 29 05:45:47 crc kubenswrapper[4799]: I1129 05:45:47.406360 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-knbck"] Nov 29 05:45:47 crc kubenswrapper[4799]: I1129 05:45:47.407111 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-knbck" podUID="6f8a509f-be52-4f76-93e5-00f5dc59f7fb" containerName="registry-server" containerID="cri-o://cebadcc259c75c7a535aebe5e9c044d01c76e8f7e2ad239e287b3c3ab5f063f8" gracePeriod=2 Nov 29 05:45:47 crc kubenswrapper[4799]: I1129 05:45:47.988762 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-knbck" Nov 29 05:45:48 crc kubenswrapper[4799]: I1129 05:45:48.036346 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xdpx6\" (UniqueName: \"kubernetes.io/projected/6f8a509f-be52-4f76-93e5-00f5dc59f7fb-kube-api-access-xdpx6\") pod \"6f8a509f-be52-4f76-93e5-00f5dc59f7fb\" (UID: \"6f8a509f-be52-4f76-93e5-00f5dc59f7fb\") " Nov 29 05:45:48 crc kubenswrapper[4799]: I1129 05:45:48.036453 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f8a509f-be52-4f76-93e5-00f5dc59f7fb-catalog-content\") pod \"6f8a509f-be52-4f76-93e5-00f5dc59f7fb\" (UID: \"6f8a509f-be52-4f76-93e5-00f5dc59f7fb\") " Nov 29 05:45:48 crc kubenswrapper[4799]: I1129 05:45:48.036487 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f8a509f-be52-4f76-93e5-00f5dc59f7fb-utilities\") pod \"6f8a509f-be52-4f76-93e5-00f5dc59f7fb\" (UID: \"6f8a509f-be52-4f76-93e5-00f5dc59f7fb\") " Nov 29 05:45:48 crc kubenswrapper[4799]: I1129 05:45:48.037582 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6f8a509f-be52-4f76-93e5-00f5dc59f7fb-utilities" (OuterVolumeSpecName: "utilities") pod "6f8a509f-be52-4f76-93e5-00f5dc59f7fb" (UID: "6f8a509f-be52-4f76-93e5-00f5dc59f7fb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:45:48 crc kubenswrapper[4799]: I1129 05:45:48.052993 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f8a509f-be52-4f76-93e5-00f5dc59f7fb-kube-api-access-xdpx6" (OuterVolumeSpecName: "kube-api-access-xdpx6") pod "6f8a509f-be52-4f76-93e5-00f5dc59f7fb" (UID: "6f8a509f-be52-4f76-93e5-00f5dc59f7fb"). InnerVolumeSpecName "kube-api-access-xdpx6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:45:48 crc kubenswrapper[4799]: I1129 05:45:48.122349 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6f8a509f-be52-4f76-93e5-00f5dc59f7fb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6f8a509f-be52-4f76-93e5-00f5dc59f7fb" (UID: "6f8a509f-be52-4f76-93e5-00f5dc59f7fb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:45:48 crc kubenswrapper[4799]: I1129 05:45:48.138117 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f8a509f-be52-4f76-93e5-00f5dc59f7fb-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 05:45:48 crc kubenswrapper[4799]: I1129 05:45:48.138213 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f8a509f-be52-4f76-93e5-00f5dc59f7fb-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 05:45:48 crc kubenswrapper[4799]: I1129 05:45:48.138267 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xdpx6\" (UniqueName: \"kubernetes.io/projected/6f8a509f-be52-4f76-93e5-00f5dc59f7fb-kube-api-access-xdpx6\") on node \"crc\" DevicePath \"\"" Nov 29 05:45:48 crc kubenswrapper[4799]: I1129 05:45:48.142138 4799 generic.go:334] "Generic (PLEG): container finished" podID="6f8a509f-be52-4f76-93e5-00f5dc59f7fb" containerID="cebadcc259c75c7a535aebe5e9c044d01c76e8f7e2ad239e287b3c3ab5f063f8" exitCode=0 Nov 29 05:45:48 crc kubenswrapper[4799]: I1129 05:45:48.142631 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-knbck" Nov 29 05:45:48 crc kubenswrapper[4799]: I1129 05:45:48.142723 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-knbck" event={"ID":"6f8a509f-be52-4f76-93e5-00f5dc59f7fb","Type":"ContainerDied","Data":"cebadcc259c75c7a535aebe5e9c044d01c76e8f7e2ad239e287b3c3ab5f063f8"} Nov 29 05:45:48 crc kubenswrapper[4799]: I1129 05:45:48.142762 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-knbck" event={"ID":"6f8a509f-be52-4f76-93e5-00f5dc59f7fb","Type":"ContainerDied","Data":"724a4426084693be9ded30332cfe8a5921be89cf719acb22d002be64f952fbff"} Nov 29 05:45:48 crc kubenswrapper[4799]: I1129 05:45:48.142804 4799 scope.go:117] "RemoveContainer" containerID="cebadcc259c75c7a535aebe5e9c044d01c76e8f7e2ad239e287b3c3ab5f063f8" Nov 29 05:45:48 crc kubenswrapper[4799]: I1129 05:45:48.188854 4799 scope.go:117] "RemoveContainer" containerID="8ad8330f282239422cacd78f34b97a74be86fb631e610218e468bf9191dcf024" Nov 29 05:45:48 crc kubenswrapper[4799]: I1129 05:45:48.220079 4799 scope.go:117] "RemoveContainer" containerID="120d5265ef50f957c208ec2016b4dc73af0337344e39abc882539af438491617" Nov 29 05:45:48 crc kubenswrapper[4799]: I1129 05:45:48.233673 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-knbck"] Nov 29 05:45:48 crc kubenswrapper[4799]: I1129 05:45:48.248271 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-knbck"] Nov 29 05:45:48 crc kubenswrapper[4799]: I1129 05:45:48.276525 4799 scope.go:117] "RemoveContainer" containerID="cebadcc259c75c7a535aebe5e9c044d01c76e8f7e2ad239e287b3c3ab5f063f8" Nov 29 05:45:48 crc kubenswrapper[4799]: E1129 05:45:48.277246 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cebadcc259c75c7a535aebe5e9c044d01c76e8f7e2ad239e287b3c3ab5f063f8\": container with ID starting with cebadcc259c75c7a535aebe5e9c044d01c76e8f7e2ad239e287b3c3ab5f063f8 not found: ID does not exist" containerID="cebadcc259c75c7a535aebe5e9c044d01c76e8f7e2ad239e287b3c3ab5f063f8" Nov 29 05:45:48 crc kubenswrapper[4799]: I1129 05:45:48.277293 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cebadcc259c75c7a535aebe5e9c044d01c76e8f7e2ad239e287b3c3ab5f063f8"} err="failed to get container status \"cebadcc259c75c7a535aebe5e9c044d01c76e8f7e2ad239e287b3c3ab5f063f8\": rpc error: code = NotFound desc = could not find container \"cebadcc259c75c7a535aebe5e9c044d01c76e8f7e2ad239e287b3c3ab5f063f8\": container with ID starting with cebadcc259c75c7a535aebe5e9c044d01c76e8f7e2ad239e287b3c3ab5f063f8 not found: ID does not exist" Nov 29 05:45:48 crc kubenswrapper[4799]: I1129 05:45:48.277328 4799 scope.go:117] "RemoveContainer" containerID="8ad8330f282239422cacd78f34b97a74be86fb631e610218e468bf9191dcf024" Nov 29 05:45:48 crc kubenswrapper[4799]: E1129 05:45:48.277593 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ad8330f282239422cacd78f34b97a74be86fb631e610218e468bf9191dcf024\": container with ID starting with 8ad8330f282239422cacd78f34b97a74be86fb631e610218e468bf9191dcf024 not found: ID does not exist" containerID="8ad8330f282239422cacd78f34b97a74be86fb631e610218e468bf9191dcf024" Nov 29 05:45:48 crc kubenswrapper[4799]: I1129 05:45:48.277616 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ad8330f282239422cacd78f34b97a74be86fb631e610218e468bf9191dcf024"} err="failed to get container status \"8ad8330f282239422cacd78f34b97a74be86fb631e610218e468bf9191dcf024\": rpc error: code = NotFound desc = could not find container \"8ad8330f282239422cacd78f34b97a74be86fb631e610218e468bf9191dcf024\": container with ID starting with 8ad8330f282239422cacd78f34b97a74be86fb631e610218e468bf9191dcf024 not found: ID does not exist" Nov 29 05:45:48 crc kubenswrapper[4799]: I1129 05:45:48.277630 4799 scope.go:117] "RemoveContainer" containerID="120d5265ef50f957c208ec2016b4dc73af0337344e39abc882539af438491617" Nov 29 05:45:48 crc kubenswrapper[4799]: E1129 05:45:48.277894 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"120d5265ef50f957c208ec2016b4dc73af0337344e39abc882539af438491617\": container with ID starting with 120d5265ef50f957c208ec2016b4dc73af0337344e39abc882539af438491617 not found: ID does not exist" containerID="120d5265ef50f957c208ec2016b4dc73af0337344e39abc882539af438491617" Nov 29 05:45:48 crc kubenswrapper[4799]: I1129 05:45:48.277913 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"120d5265ef50f957c208ec2016b4dc73af0337344e39abc882539af438491617"} err="failed to get container status \"120d5265ef50f957c208ec2016b4dc73af0337344e39abc882539af438491617\": rpc error: code = NotFound desc = could not find container \"120d5265ef50f957c208ec2016b4dc73af0337344e39abc882539af438491617\": container with ID starting with 120d5265ef50f957c208ec2016b4dc73af0337344e39abc882539af438491617 not found: ID does not exist" Nov 29 05:45:48 crc kubenswrapper[4799]: I1129 05:45:48.670919 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f8a509f-be52-4f76-93e5-00f5dc59f7fb" path="/var/lib/kubelet/pods/6f8a509f-be52-4f76-93e5-00f5dc59f7fb/volumes" Nov 29 05:45:59 crc kubenswrapper[4799]: I1129 05:45:59.114262 4799 scope.go:117] "RemoveContainer" containerID="f038d0ded959a6574153aa96512043f347f81bf952618e47406a40b4a8b7221b" Nov 29 05:46:07 crc kubenswrapper[4799]: I1129 05:46:07.497783 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 05:46:07 crc kubenswrapper[4799]: I1129 05:46:07.498727 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 05:46:15 crc kubenswrapper[4799]: I1129 05:46:15.407809 4799 generic.go:334] "Generic (PLEG): container finished" podID="ac2f0929-7fb0-4e7e-b692-e3f1da437287" containerID="ce12a605982e76f68a3be954db1c235af8e583a9c7ae7885629e30d988a8a4c3" exitCode=0 Nov 29 05:46:15 crc kubenswrapper[4799]: I1129 05:46:15.407901 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ll7z5/crc-debug-j2w64" event={"ID":"ac2f0929-7fb0-4e7e-b692-e3f1da437287","Type":"ContainerDied","Data":"ce12a605982e76f68a3be954db1c235af8e583a9c7ae7885629e30d988a8a4c3"} Nov 29 05:46:16 crc kubenswrapper[4799]: I1129 05:46:16.530208 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ll7z5/crc-debug-j2w64" Nov 29 05:46:16 crc kubenswrapper[4799]: I1129 05:46:16.571196 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-ll7z5/crc-debug-j2w64"] Nov 29 05:46:16 crc kubenswrapper[4799]: I1129 05:46:16.581354 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-ll7z5/crc-debug-j2w64"] Nov 29 05:46:16 crc kubenswrapper[4799]: I1129 05:46:16.732450 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxngm\" (UniqueName: \"kubernetes.io/projected/ac2f0929-7fb0-4e7e-b692-e3f1da437287-kube-api-access-wxngm\") pod \"ac2f0929-7fb0-4e7e-b692-e3f1da437287\" (UID: \"ac2f0929-7fb0-4e7e-b692-e3f1da437287\") " Nov 29 05:46:16 crc kubenswrapper[4799]: I1129 05:46:16.732717 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ac2f0929-7fb0-4e7e-b692-e3f1da437287-host\") pod \"ac2f0929-7fb0-4e7e-b692-e3f1da437287\" (UID: \"ac2f0929-7fb0-4e7e-b692-e3f1da437287\") " Nov 29 05:46:16 crc kubenswrapper[4799]: I1129 05:46:16.733033 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ac2f0929-7fb0-4e7e-b692-e3f1da437287-host" (OuterVolumeSpecName: "host") pod "ac2f0929-7fb0-4e7e-b692-e3f1da437287" (UID: "ac2f0929-7fb0-4e7e-b692-e3f1da437287"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 05:46:16 crc kubenswrapper[4799]: I1129 05:46:16.734029 4799 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ac2f0929-7fb0-4e7e-b692-e3f1da437287-host\") on node \"crc\" DevicePath \"\"" Nov 29 05:46:16 crc kubenswrapper[4799]: I1129 05:46:16.741033 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac2f0929-7fb0-4e7e-b692-e3f1da437287-kube-api-access-wxngm" (OuterVolumeSpecName: "kube-api-access-wxngm") pod "ac2f0929-7fb0-4e7e-b692-e3f1da437287" (UID: "ac2f0929-7fb0-4e7e-b692-e3f1da437287"). InnerVolumeSpecName "kube-api-access-wxngm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:46:16 crc kubenswrapper[4799]: I1129 05:46:16.835836 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxngm\" (UniqueName: \"kubernetes.io/projected/ac2f0929-7fb0-4e7e-b692-e3f1da437287-kube-api-access-wxngm\") on node \"crc\" DevicePath \"\"" Nov 29 05:46:17 crc kubenswrapper[4799]: I1129 05:46:17.427568 4799 scope.go:117] "RemoveContainer" containerID="ce12a605982e76f68a3be954db1c235af8e583a9c7ae7885629e30d988a8a4c3" Nov 29 05:46:17 crc kubenswrapper[4799]: I1129 05:46:17.427696 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ll7z5/crc-debug-j2w64" Nov 29 05:46:17 crc kubenswrapper[4799]: I1129 05:46:17.732926 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-ll7z5/crc-debug-gz875"] Nov 29 05:46:17 crc kubenswrapper[4799]: E1129 05:46:17.733666 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f8a509f-be52-4f76-93e5-00f5dc59f7fb" containerName="extract-utilities" Nov 29 05:46:17 crc kubenswrapper[4799]: I1129 05:46:17.733681 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f8a509f-be52-4f76-93e5-00f5dc59f7fb" containerName="extract-utilities" Nov 29 05:46:17 crc kubenswrapper[4799]: E1129 05:46:17.733708 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f8a509f-be52-4f76-93e5-00f5dc59f7fb" containerName="registry-server" Nov 29 05:46:17 crc kubenswrapper[4799]: I1129 05:46:17.733716 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f8a509f-be52-4f76-93e5-00f5dc59f7fb" containerName="registry-server" Nov 29 05:46:17 crc kubenswrapper[4799]: E1129 05:46:17.733724 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac2f0929-7fb0-4e7e-b692-e3f1da437287" containerName="container-00" Nov 29 05:46:17 crc kubenswrapper[4799]: I1129 05:46:17.733730 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac2f0929-7fb0-4e7e-b692-e3f1da437287" containerName="container-00" Nov 29 05:46:17 crc kubenswrapper[4799]: E1129 05:46:17.733746 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f8a509f-be52-4f76-93e5-00f5dc59f7fb" containerName="extract-content" Nov 29 05:46:17 crc kubenswrapper[4799]: I1129 05:46:17.733752 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f8a509f-be52-4f76-93e5-00f5dc59f7fb" containerName="extract-content" Nov 29 05:46:17 crc kubenswrapper[4799]: I1129 05:46:17.733948 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac2f0929-7fb0-4e7e-b692-e3f1da437287" containerName="container-00" Nov 29 05:46:17 crc kubenswrapper[4799]: I1129 05:46:17.733968 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f8a509f-be52-4f76-93e5-00f5dc59f7fb" containerName="registry-server" Nov 29 05:46:17 crc kubenswrapper[4799]: I1129 05:46:17.734840 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ll7z5/crc-debug-gz875" Nov 29 05:46:17 crc kubenswrapper[4799]: I1129 05:46:17.737100 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-ll7z5"/"default-dockercfg-h574n" Nov 29 05:46:17 crc kubenswrapper[4799]: I1129 05:46:17.753583 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/81dadd2f-f6d6-4951-94b5-b0d3b14012c6-host\") pod \"crc-debug-gz875\" (UID: \"81dadd2f-f6d6-4951-94b5-b0d3b14012c6\") " pod="openshift-must-gather-ll7z5/crc-debug-gz875" Nov 29 05:46:17 crc kubenswrapper[4799]: I1129 05:46:17.753680 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfxfq\" (UniqueName: \"kubernetes.io/projected/81dadd2f-f6d6-4951-94b5-b0d3b14012c6-kube-api-access-kfxfq\") pod \"crc-debug-gz875\" (UID: \"81dadd2f-f6d6-4951-94b5-b0d3b14012c6\") " pod="openshift-must-gather-ll7z5/crc-debug-gz875" Nov 29 05:46:17 crc kubenswrapper[4799]: I1129 05:46:17.855555 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/81dadd2f-f6d6-4951-94b5-b0d3b14012c6-host\") pod \"crc-debug-gz875\" (UID: \"81dadd2f-f6d6-4951-94b5-b0d3b14012c6\") " pod="openshift-must-gather-ll7z5/crc-debug-gz875" Nov 29 05:46:17 crc kubenswrapper[4799]: I1129 05:46:17.855628 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfxfq\" (UniqueName: \"kubernetes.io/projected/81dadd2f-f6d6-4951-94b5-b0d3b14012c6-kube-api-access-kfxfq\") pod \"crc-debug-gz875\" (UID: \"81dadd2f-f6d6-4951-94b5-b0d3b14012c6\") " pod="openshift-must-gather-ll7z5/crc-debug-gz875" Nov 29 05:46:17 crc kubenswrapper[4799]: I1129 05:46:17.856102 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/81dadd2f-f6d6-4951-94b5-b0d3b14012c6-host\") pod \"crc-debug-gz875\" (UID: \"81dadd2f-f6d6-4951-94b5-b0d3b14012c6\") " pod="openshift-must-gather-ll7z5/crc-debug-gz875" Nov 29 05:46:17 crc kubenswrapper[4799]: I1129 05:46:17.898843 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfxfq\" (UniqueName: \"kubernetes.io/projected/81dadd2f-f6d6-4951-94b5-b0d3b14012c6-kube-api-access-kfxfq\") pod \"crc-debug-gz875\" (UID: \"81dadd2f-f6d6-4951-94b5-b0d3b14012c6\") " pod="openshift-must-gather-ll7z5/crc-debug-gz875" Nov 29 05:46:18 crc kubenswrapper[4799]: I1129 05:46:18.051301 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ll7z5/crc-debug-gz875" Nov 29 05:46:18 crc kubenswrapper[4799]: I1129 05:46:18.437960 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ll7z5/crc-debug-gz875" event={"ID":"81dadd2f-f6d6-4951-94b5-b0d3b14012c6","Type":"ContainerStarted","Data":"acb221c71e84f6bbe7acb640190a9cfc97779df98a8cf11bbac8de2e43b0a176"} Nov 29 05:46:18 crc kubenswrapper[4799]: I1129 05:46:18.668962 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac2f0929-7fb0-4e7e-b692-e3f1da437287" path="/var/lib/kubelet/pods/ac2f0929-7fb0-4e7e-b692-e3f1da437287/volumes" Nov 29 05:46:19 crc kubenswrapper[4799]: I1129 05:46:19.448211 4799 generic.go:334] "Generic (PLEG): container finished" podID="81dadd2f-f6d6-4951-94b5-b0d3b14012c6" containerID="f7453ef5077138b52ed9b6487e7eb3d2cc030232cabb278103bcf9d493387c34" exitCode=0 Nov 29 05:46:19 crc kubenswrapper[4799]: I1129 05:46:19.448304 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ll7z5/crc-debug-gz875" event={"ID":"81dadd2f-f6d6-4951-94b5-b0d3b14012c6","Type":"ContainerDied","Data":"f7453ef5077138b52ed9b6487e7eb3d2cc030232cabb278103bcf9d493387c34"} Nov 29 05:46:20 crc kubenswrapper[4799]: I1129 05:46:20.545672 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ll7z5/crc-debug-gz875" Nov 29 05:46:20 crc kubenswrapper[4799]: I1129 05:46:20.708429 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfxfq\" (UniqueName: \"kubernetes.io/projected/81dadd2f-f6d6-4951-94b5-b0d3b14012c6-kube-api-access-kfxfq\") pod \"81dadd2f-f6d6-4951-94b5-b0d3b14012c6\" (UID: \"81dadd2f-f6d6-4951-94b5-b0d3b14012c6\") " Nov 29 05:46:20 crc kubenswrapper[4799]: I1129 05:46:20.708522 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/81dadd2f-f6d6-4951-94b5-b0d3b14012c6-host\") pod \"81dadd2f-f6d6-4951-94b5-b0d3b14012c6\" (UID: \"81dadd2f-f6d6-4951-94b5-b0d3b14012c6\") " Nov 29 05:46:20 crc kubenswrapper[4799]: I1129 05:46:20.708852 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/81dadd2f-f6d6-4951-94b5-b0d3b14012c6-host" (OuterVolumeSpecName: "host") pod "81dadd2f-f6d6-4951-94b5-b0d3b14012c6" (UID: "81dadd2f-f6d6-4951-94b5-b0d3b14012c6"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 05:46:20 crc kubenswrapper[4799]: I1129 05:46:20.709008 4799 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/81dadd2f-f6d6-4951-94b5-b0d3b14012c6-host\") on node \"crc\" DevicePath \"\"" Nov 29 05:46:20 crc kubenswrapper[4799]: I1129 05:46:20.719219 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81dadd2f-f6d6-4951-94b5-b0d3b14012c6-kube-api-access-kfxfq" (OuterVolumeSpecName: "kube-api-access-kfxfq") pod "81dadd2f-f6d6-4951-94b5-b0d3b14012c6" (UID: "81dadd2f-f6d6-4951-94b5-b0d3b14012c6"). InnerVolumeSpecName "kube-api-access-kfxfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:46:20 crc kubenswrapper[4799]: I1129 05:46:20.810505 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfxfq\" (UniqueName: \"kubernetes.io/projected/81dadd2f-f6d6-4951-94b5-b0d3b14012c6-kube-api-access-kfxfq\") on node \"crc\" DevicePath \"\"" Nov 29 05:46:21 crc kubenswrapper[4799]: I1129 05:46:21.470393 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ll7z5/crc-debug-gz875" event={"ID":"81dadd2f-f6d6-4951-94b5-b0d3b14012c6","Type":"ContainerDied","Data":"acb221c71e84f6bbe7acb640190a9cfc97779df98a8cf11bbac8de2e43b0a176"} Nov 29 05:46:21 crc kubenswrapper[4799]: I1129 05:46:21.470729 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="acb221c71e84f6bbe7acb640190a9cfc97779df98a8cf11bbac8de2e43b0a176" Nov 29 05:46:21 crc kubenswrapper[4799]: I1129 05:46:21.470857 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ll7z5/crc-debug-gz875" Nov 29 05:46:21 crc kubenswrapper[4799]: I1129 05:46:21.473606 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-ll7z5/crc-debug-gz875"] Nov 29 05:46:21 crc kubenswrapper[4799]: I1129 05:46:21.484090 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-ll7z5/crc-debug-gz875"] Nov 29 05:46:22 crc kubenswrapper[4799]: I1129 05:46:22.671945 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81dadd2f-f6d6-4951-94b5-b0d3b14012c6" path="/var/lib/kubelet/pods/81dadd2f-f6d6-4951-94b5-b0d3b14012c6/volumes" Nov 29 05:46:22 crc kubenswrapper[4799]: I1129 05:46:22.672621 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-ll7z5/crc-debug-gcsqm"] Nov 29 05:46:22 crc kubenswrapper[4799]: E1129 05:46:22.672990 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81dadd2f-f6d6-4951-94b5-b0d3b14012c6" containerName="container-00" Nov 29 05:46:22 crc kubenswrapper[4799]: I1129 05:46:22.673006 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="81dadd2f-f6d6-4951-94b5-b0d3b14012c6" containerName="container-00" Nov 29 05:46:22 crc kubenswrapper[4799]: I1129 05:46:22.673660 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="81dadd2f-f6d6-4951-94b5-b0d3b14012c6" containerName="container-00" Nov 29 05:46:22 crc kubenswrapper[4799]: I1129 05:46:22.674462 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ll7z5/crc-debug-gcsqm" Nov 29 05:46:22 crc kubenswrapper[4799]: I1129 05:46:22.677542 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-ll7z5"/"default-dockercfg-h574n" Nov 29 05:46:22 crc kubenswrapper[4799]: I1129 05:46:22.748472 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/147b56fb-2ad7-4807-b27b-de550b69cfda-host\") pod \"crc-debug-gcsqm\" (UID: \"147b56fb-2ad7-4807-b27b-de550b69cfda\") " pod="openshift-must-gather-ll7z5/crc-debug-gcsqm" Nov 29 05:46:22 crc kubenswrapper[4799]: I1129 05:46:22.748583 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgcwl\" (UniqueName: \"kubernetes.io/projected/147b56fb-2ad7-4807-b27b-de550b69cfda-kube-api-access-kgcwl\") pod \"crc-debug-gcsqm\" (UID: \"147b56fb-2ad7-4807-b27b-de550b69cfda\") " pod="openshift-must-gather-ll7z5/crc-debug-gcsqm" Nov 29 05:46:22 crc kubenswrapper[4799]: I1129 05:46:22.851498 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/147b56fb-2ad7-4807-b27b-de550b69cfda-host\") pod \"crc-debug-gcsqm\" (UID: \"147b56fb-2ad7-4807-b27b-de550b69cfda\") " pod="openshift-must-gather-ll7z5/crc-debug-gcsqm" Nov 29 05:46:22 crc kubenswrapper[4799]: I1129 05:46:22.851572 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgcwl\" (UniqueName: \"kubernetes.io/projected/147b56fb-2ad7-4807-b27b-de550b69cfda-kube-api-access-kgcwl\") pod \"crc-debug-gcsqm\" (UID: \"147b56fb-2ad7-4807-b27b-de550b69cfda\") " pod="openshift-must-gather-ll7z5/crc-debug-gcsqm" Nov 29 05:46:22 crc kubenswrapper[4799]: I1129 05:46:22.851682 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/147b56fb-2ad7-4807-b27b-de550b69cfda-host\") pod \"crc-debug-gcsqm\" (UID: \"147b56fb-2ad7-4807-b27b-de550b69cfda\") " pod="openshift-must-gather-ll7z5/crc-debug-gcsqm" Nov 29 05:46:22 crc kubenswrapper[4799]: I1129 05:46:22.874420 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgcwl\" (UniqueName: \"kubernetes.io/projected/147b56fb-2ad7-4807-b27b-de550b69cfda-kube-api-access-kgcwl\") pod \"crc-debug-gcsqm\" (UID: \"147b56fb-2ad7-4807-b27b-de550b69cfda\") " pod="openshift-must-gather-ll7z5/crc-debug-gcsqm" Nov 29 05:46:22 crc kubenswrapper[4799]: I1129 05:46:22.994639 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ll7z5/crc-debug-gcsqm" Nov 29 05:46:23 crc kubenswrapper[4799]: W1129 05:46:23.032278 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod147b56fb_2ad7_4807_b27b_de550b69cfda.slice/crio-da592e3a4f729c58bdf1f63374b3dc45b88ecbf0093dda6f040137b27089c3ac WatchSource:0}: Error finding container da592e3a4f729c58bdf1f63374b3dc45b88ecbf0093dda6f040137b27089c3ac: Status 404 returned error can't find the container with id da592e3a4f729c58bdf1f63374b3dc45b88ecbf0093dda6f040137b27089c3ac Nov 29 05:46:23 crc kubenswrapper[4799]: I1129 05:46:23.490354 4799 generic.go:334] "Generic (PLEG): container finished" podID="147b56fb-2ad7-4807-b27b-de550b69cfda" containerID="f8f4854c1cd95b0e6a3ad284a8045fe7db54d85552349f6b16b37eaa63d0216f" exitCode=0 Nov 29 05:46:23 crc kubenswrapper[4799]: I1129 05:46:23.490408 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ll7z5/crc-debug-gcsqm" event={"ID":"147b56fb-2ad7-4807-b27b-de550b69cfda","Type":"ContainerDied","Data":"f8f4854c1cd95b0e6a3ad284a8045fe7db54d85552349f6b16b37eaa63d0216f"} Nov 29 05:46:23 crc kubenswrapper[4799]: I1129 05:46:23.491036 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ll7z5/crc-debug-gcsqm" event={"ID":"147b56fb-2ad7-4807-b27b-de550b69cfda","Type":"ContainerStarted","Data":"da592e3a4f729c58bdf1f63374b3dc45b88ecbf0093dda6f040137b27089c3ac"} Nov 29 05:46:23 crc kubenswrapper[4799]: I1129 05:46:23.535514 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-ll7z5/crc-debug-gcsqm"] Nov 29 05:46:23 crc kubenswrapper[4799]: I1129 05:46:23.548005 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-ll7z5/crc-debug-gcsqm"] Nov 29 05:46:25 crc kubenswrapper[4799]: I1129 05:46:25.041569 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ll7z5/crc-debug-gcsqm" Nov 29 05:46:25 crc kubenswrapper[4799]: I1129 05:46:25.111222 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kgcwl\" (UniqueName: \"kubernetes.io/projected/147b56fb-2ad7-4807-b27b-de550b69cfda-kube-api-access-kgcwl\") pod \"147b56fb-2ad7-4807-b27b-de550b69cfda\" (UID: \"147b56fb-2ad7-4807-b27b-de550b69cfda\") " Nov 29 05:46:25 crc kubenswrapper[4799]: I1129 05:46:25.111335 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/147b56fb-2ad7-4807-b27b-de550b69cfda-host\") pod \"147b56fb-2ad7-4807-b27b-de550b69cfda\" (UID: \"147b56fb-2ad7-4807-b27b-de550b69cfda\") " Nov 29 05:46:25 crc kubenswrapper[4799]: I1129 05:46:25.111521 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/147b56fb-2ad7-4807-b27b-de550b69cfda-host" (OuterVolumeSpecName: "host") pod "147b56fb-2ad7-4807-b27b-de550b69cfda" (UID: "147b56fb-2ad7-4807-b27b-de550b69cfda"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 05:46:25 crc kubenswrapper[4799]: I1129 05:46:25.112384 4799 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/147b56fb-2ad7-4807-b27b-de550b69cfda-host\") on node \"crc\" DevicePath \"\"" Nov 29 05:46:25 crc kubenswrapper[4799]: I1129 05:46:25.120278 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/147b56fb-2ad7-4807-b27b-de550b69cfda-kube-api-access-kgcwl" (OuterVolumeSpecName: "kube-api-access-kgcwl") pod "147b56fb-2ad7-4807-b27b-de550b69cfda" (UID: "147b56fb-2ad7-4807-b27b-de550b69cfda"). InnerVolumeSpecName "kube-api-access-kgcwl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:46:25 crc kubenswrapper[4799]: I1129 05:46:25.216343 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kgcwl\" (UniqueName: \"kubernetes.io/projected/147b56fb-2ad7-4807-b27b-de550b69cfda-kube-api-access-kgcwl\") on node \"crc\" DevicePath \"\"" Nov 29 05:46:25 crc kubenswrapper[4799]: I1129 05:46:25.509694 4799 scope.go:117] "RemoveContainer" containerID="f8f4854c1cd95b0e6a3ad284a8045fe7db54d85552349f6b16b37eaa63d0216f" Nov 29 05:46:25 crc kubenswrapper[4799]: I1129 05:46:25.509733 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ll7z5/crc-debug-gcsqm" Nov 29 05:46:26 crc kubenswrapper[4799]: I1129 05:46:26.671647 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="147b56fb-2ad7-4807-b27b-de550b69cfda" path="/var/lib/kubelet/pods/147b56fb-2ad7-4807-b27b-de550b69cfda/volumes" Nov 29 05:46:37 crc kubenswrapper[4799]: I1129 05:46:37.497495 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 05:46:37 crc kubenswrapper[4799]: I1129 05:46:37.499162 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 05:46:47 crc kubenswrapper[4799]: I1129 05:46:47.253534 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-fdfbddb56-lhh8r_ddd039b3-89fb-4515-a633-e4c8c23bc25c/barbican-api/0.log" Nov 29 05:46:47 crc kubenswrapper[4799]: I1129 05:46:47.388353 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-fdfbddb56-lhh8r_ddd039b3-89fb-4515-a633-e4c8c23bc25c/barbican-api-log/0.log" Nov 29 05:46:47 crc kubenswrapper[4799]: I1129 05:46:47.512483 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6495665bf6-xn8f7_bdd214a8-9094-42ef-b2c6-2452b59596f7/barbican-keystone-listener/0.log" Nov 29 05:46:47 crc kubenswrapper[4799]: I1129 05:46:47.735516 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6495665bf6-xn8f7_bdd214a8-9094-42ef-b2c6-2452b59596f7/barbican-keystone-listener-log/0.log" Nov 29 05:46:47 crc kubenswrapper[4799]: I1129 05:46:47.756024 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-76859cbf9f-kdxtl_f14d7038-795e-4f8a-8ce7-81e64c86137b/barbican-worker/0.log" Nov 29 05:46:47 crc kubenswrapper[4799]: I1129 05:46:47.804129 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-76859cbf9f-kdxtl_f14d7038-795e-4f8a-8ce7-81e64c86137b/barbican-worker-log/0.log" Nov 29 05:46:48 crc kubenswrapper[4799]: I1129 05:46:48.006591 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-k9rvx_a30efa1d-c61a-4bc1-9350-fa1059e7bb71/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 05:46:48 crc kubenswrapper[4799]: I1129 05:46:48.016051 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c/ceilometer-central-agent/0.log" Nov 29 05:46:48 crc kubenswrapper[4799]: I1129 05:46:48.138582 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c/ceilometer-notification-agent/0.log" Nov 29 05:46:48 crc kubenswrapper[4799]: I1129 05:46:48.217309 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c/proxy-httpd/0.log" Nov 29 05:46:48 crc kubenswrapper[4799]: I1129 05:46:48.223403 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c/sg-core/0.log" Nov 29 05:46:48 crc kubenswrapper[4799]: I1129 05:46:48.338358 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-edpm-deployment-openstack-edpm-ipam-fdll7_41294840-0c22-4c85-ae8f-31f3a00458d4/ceph-client-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 05:46:48 crc kubenswrapper[4799]: I1129 05:46:48.476206 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6rm8k_92efc83b-d2e4-44f7-8ede-db45686a6c0f/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 05:46:48 crc kubenswrapper[4799]: I1129 05:46:48.613177 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_64042cd5-9bc5-4371-b063-753befa911c2/cinder-api/0.log" Nov 29 05:46:48 crc kubenswrapper[4799]: I1129 05:46:48.666294 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_64042cd5-9bc5-4371-b063-753befa911c2/cinder-api-log/0.log" Nov 29 05:46:48 crc kubenswrapper[4799]: I1129 05:46:48.845958 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_caa349ef-0de7-46c4-a7cd-fdf43aa6f949/cinder-backup/0.log" Nov 29 05:46:48 crc kubenswrapper[4799]: I1129 05:46:48.898624 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_caa349ef-0de7-46c4-a7cd-fdf43aa6f949/probe/0.log" Nov 29 05:46:49 crc kubenswrapper[4799]: I1129 05:46:49.277648 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_6c7456e1-235c-4bb3-a587-f7e3c620749e/cinder-scheduler/0.log" Nov 29 05:46:49 crc kubenswrapper[4799]: I1129 05:46:49.284561 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_6c7456e1-235c-4bb3-a587-f7e3c620749e/probe/0.log" Nov 29 05:46:49 crc kubenswrapper[4799]: I1129 05:46:49.382433 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_0907967b-0bf5-4fe0-9862-ffbd54efde7f/cinder-volume/0.log" Nov 29 05:46:49 crc kubenswrapper[4799]: I1129 05:46:49.487845 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_0907967b-0bf5-4fe0-9862-ffbd54efde7f/probe/0.log" Nov 29 05:46:49 crc kubenswrapper[4799]: I1129 05:46:49.644270 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-x4w52_86cf64fe-3c75-4b05-8503-a4e3f3e0395c/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 05:46:49 crc kubenswrapper[4799]: I1129 05:46:49.757883 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-k7vvz_56293cbb-ed95-4541-b322-8e86233a8ff5/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 05:46:49 crc kubenswrapper[4799]: I1129 05:46:49.862336 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-76b5fdb995-cbg8h_87fc92de-e3e3-454f-886a-226aa591c580/init/0.log" Nov 29 05:46:50 crc kubenswrapper[4799]: I1129 05:46:50.030695 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-76b5fdb995-cbg8h_87fc92de-e3e3-454f-886a-226aa591c580/init/0.log" Nov 29 05:46:50 crc kubenswrapper[4799]: I1129 05:46:50.090894 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-76b5fdb995-cbg8h_87fc92de-e3e3-454f-886a-226aa591c580/dnsmasq-dns/0.log" Nov 29 05:46:50 crc kubenswrapper[4799]: I1129 05:46:50.109670 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_cd7d046d-4862-47c2-b4d3-e4d81d79706a/glance-httpd/0.log" Nov 29 05:46:50 crc kubenswrapper[4799]: I1129 05:46:50.294704 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_cd7d046d-4862-47c2-b4d3-e4d81d79706a/glance-log/0.log" Nov 29 05:46:51 crc kubenswrapper[4799]: I1129 05:46:51.584321 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_477f0d92-5648-405b-8034-f851200268a9/glance-log/0.log" Nov 29 05:46:51 crc kubenswrapper[4799]: I1129 05:46:51.658671 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7c5488f496-jvw7p_93fb8bea-4246-40da-b480-c980b628e2b9/horizon/0.log" Nov 29 05:46:51 crc kubenswrapper[4799]: I1129 05:46:51.691964 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_477f0d92-5648-405b-8034-f851200268a9/glance-httpd/0.log" Nov 29 05:46:51 crc kubenswrapper[4799]: I1129 05:46:51.837057 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv_cbad88f8-b661-4b9f-8815-6c0f2df63d7f/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 05:46:51 crc kubenswrapper[4799]: I1129 05:46:51.859926 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7c5488f496-jvw7p_93fb8bea-4246-40da-b480-c980b628e2b9/horizon-log/0.log" Nov 29 05:46:52 crc kubenswrapper[4799]: I1129 05:46:52.752939 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-hdkhl_1784d248-9169-4b88-b8d1-16412106a8dc/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 05:46:52 crc kubenswrapper[4799]: I1129 05:46:52.778175 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29406541-dvcxf_4b4eb06c-ec4a-45f3-b316-34a4410ea46b/keystone-cron/0.log" Nov 29 05:46:52 crc kubenswrapper[4799]: I1129 05:46:52.942779 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_66c61a7f-a655-46e9-8a85-de7ad52bfe6f/kube-state-metrics/0.log" Nov 29 05:46:53 crc kubenswrapper[4799]: I1129 05:46:53.049816 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-v72s9_168cfe40-080e-44fa-863e-8eb8b5b5923c/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 05:46:53 crc kubenswrapper[4799]: I1129 05:46:53.470353 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_b8b16394-158d-47e0-b406-18636e57be57/manila-api/0.log" Nov 29 05:46:53 crc kubenswrapper[4799]: I1129 05:46:53.479915 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_ef775c01-8e04-4efe-8192-889ddf0ccbd3/probe/0.log" Nov 29 05:46:53 crc kubenswrapper[4799]: I1129 05:46:53.503845 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-757bb6ddd6-9wg88_5d5a9855-0646-45ff-8036-afba2a0d1a41/keystone-api/0.log" Nov 29 05:46:53 crc kubenswrapper[4799]: I1129 05:46:53.707129 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_ef775c01-8e04-4efe-8192-889ddf0ccbd3/manila-scheduler/0.log" Nov 29 05:46:53 crc kubenswrapper[4799]: I1129 05:46:53.768289 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_d858c44e-0d42-4c54-b820-dc89173d67d1/probe/0.log" Nov 29 05:46:53 crc kubenswrapper[4799]: I1129 05:46:53.957883 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_d858c44e-0d42-4c54-b820-dc89173d67d1/manila-share/0.log" Nov 29 05:46:54 crc kubenswrapper[4799]: I1129 05:46:54.071715 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_b8b16394-158d-47e0-b406-18636e57be57/manila-api-log/0.log" Nov 29 05:46:54 crc kubenswrapper[4799]: I1129 05:46:54.314235 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-669f887b5-74klb_122be3fa-bc40-45f0-808d-fece3e876c43/neutron-api/0.log" Nov 29 05:46:54 crc kubenswrapper[4799]: I1129 05:46:54.333305 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-669f887b5-74klb_122be3fa-bc40-45f0-808d-fece3e876c43/neutron-httpd/0.log" Nov 29 05:46:54 crc kubenswrapper[4799]: I1129 05:46:54.540735 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9_8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 05:46:54 crc kubenswrapper[4799]: I1129 05:46:54.934565 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_cd1dc76e-76ba-4ab4-b40c-587ffbf9b5df/nova-api-log/0.log" Nov 29 05:46:55 crc kubenswrapper[4799]: I1129 05:46:55.005445 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_f6a2c6d8-ac5d-4b14-a907-d624f22d5f1b/nova-cell0-conductor-conductor/0.log" Nov 29 05:46:55 crc kubenswrapper[4799]: I1129 05:46:55.089419 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_cd1dc76e-76ba-4ab4-b40c-587ffbf9b5df/nova-api-api/0.log" Nov 29 05:46:55 crc kubenswrapper[4799]: I1129 05:46:55.393362 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_27b8e913-bbd1-4a53-ab29-9b31ab8c64c8/nova-cell1-conductor-conductor/0.log" Nov 29 05:46:55 crc kubenswrapper[4799]: I1129 05:46:55.396508 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_91d09977-e990-4df4-921c-6f087e18b85f/nova-cell1-novncproxy-novncproxy/0.log" Nov 29 05:46:55 crc kubenswrapper[4799]: I1129 05:46:55.595136 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b_26a0b212-c652-4475-8210-4b62b653ed79/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 05:46:55 crc kubenswrapper[4799]: I1129 05:46:55.648458 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_bfbbd39d-1e8f-4bdd-991e-71e206bb4ca0/nova-metadata-log/0.log" Nov 29 05:46:55 crc kubenswrapper[4799]: I1129 05:46:55.950516 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_923b93cd-9a7f-4924-8d46-a1cd33612af5/nova-scheduler-scheduler/0.log" Nov 29 05:46:56 crc kubenswrapper[4799]: I1129 05:46:56.040205 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_735902f6-652e-4c85-818d-10f7d9529788/mysql-bootstrap/0.log" Nov 29 05:46:56 crc kubenswrapper[4799]: I1129 05:46:56.193082 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_735902f6-652e-4c85-818d-10f7d9529788/mysql-bootstrap/0.log" Nov 29 05:46:56 crc kubenswrapper[4799]: I1129 05:46:56.229612 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_735902f6-652e-4c85-818d-10f7d9529788/galera/0.log" Nov 29 05:46:56 crc kubenswrapper[4799]: I1129 05:46:56.373415 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_bbc1f3a5-585d-4e6d-a87e-cf77a7b21147/mysql-bootstrap/0.log" Nov 29 05:46:56 crc kubenswrapper[4799]: I1129 05:46:56.683406 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_bbc1f3a5-585d-4e6d-a87e-cf77a7b21147/mysql-bootstrap/0.log" Nov 29 05:46:56 crc kubenswrapper[4799]: I1129 05:46:56.707695 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_bbc1f3a5-585d-4e6d-a87e-cf77a7b21147/galera/0.log" Nov 29 05:46:56 crc kubenswrapper[4799]: I1129 05:46:56.808845 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_548ff572-b2d0-45ec-a131-0ba87044b4ff/openstackclient/0.log" Nov 29 05:46:56 crc kubenswrapper[4799]: I1129 05:46:56.957914 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-kkdw8_1d3d81e5-591a-469d-9851-86271162d455/ovn-controller/0.log" Nov 29 05:46:57 crc kubenswrapper[4799]: I1129 05:46:57.083728 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_bfbbd39d-1e8f-4bdd-991e-71e206bb4ca0/nova-metadata-metadata/0.log" Nov 29 05:46:57 crc kubenswrapper[4799]: I1129 05:46:57.145893 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-5szsr_b0d261b9-8134-46ec-b510-63fce197ccff/openstack-network-exporter/0.log" Nov 29 05:46:57 crc kubenswrapper[4799]: I1129 05:46:57.302906 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-bkkst_e358b742-108b-4f60-989f-2e8eddce487e/ovsdb-server-init/0.log" Nov 29 05:46:57 crc kubenswrapper[4799]: I1129 05:46:57.457364 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-bkkst_e358b742-108b-4f60-989f-2e8eddce487e/ovsdb-server/0.log" Nov 29 05:46:57 crc kubenswrapper[4799]: I1129 05:46:57.503667 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-bkkst_e358b742-108b-4f60-989f-2e8eddce487e/ovsdb-server-init/0.log" Nov 29 05:46:57 crc kubenswrapper[4799]: I1129 05:46:57.508576 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-bkkst_e358b742-108b-4f60-989f-2e8eddce487e/ovs-vswitchd/0.log" Nov 29 05:46:57 crc kubenswrapper[4799]: I1129 05:46:57.738085 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-hnnmv_a9783465-b33b-428e-9691-bc9b341e37ad/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 05:46:57 crc kubenswrapper[4799]: I1129 05:46:57.738111 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_c8c0babb-3163-4a72-a6d5-4fc5e64bee31/openstack-network-exporter/0.log" Nov 29 05:46:57 crc kubenswrapper[4799]: I1129 05:46:57.816438 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_c8c0babb-3163-4a72-a6d5-4fc5e64bee31/ovn-northd/0.log" Nov 29 05:46:58 crc kubenswrapper[4799]: I1129 05:46:58.004056 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_c79e4fc0-8786-40a6-a775-29a4767d072e/openstack-network-exporter/0.log" Nov 29 05:46:58 crc kubenswrapper[4799]: I1129 05:46:58.072703 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_c79e4fc0-8786-40a6-a775-29a4767d072e/ovsdbserver-nb/0.log" Nov 29 05:46:58 crc kubenswrapper[4799]: I1129 05:46:58.220237 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_5bb0e45f-f281-49ec-8a11-37e1ace553db/openstack-network-exporter/0.log" Nov 29 05:46:58 crc kubenswrapper[4799]: I1129 05:46:58.575586 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_5bb0e45f-f281-49ec-8a11-37e1ace553db/ovsdbserver-sb/0.log" Nov 29 05:46:58 crc kubenswrapper[4799]: I1129 05:46:58.682046 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-dc694ffdd-sm2s5_8af48d85-03bf-4d5f-a677-49bc96f0dfee/placement-api/0.log" Nov 29 05:46:58 crc kubenswrapper[4799]: I1129 05:46:58.795642 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-dc694ffdd-sm2s5_8af48d85-03bf-4d5f-a677-49bc96f0dfee/placement-log/0.log" Nov 29 05:46:58 crc kubenswrapper[4799]: I1129 05:46:58.839356 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac/setup-container/0.log" Nov 29 05:46:59 crc kubenswrapper[4799]: I1129 05:46:59.021717 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac/setup-container/0.log" Nov 29 05:46:59 crc kubenswrapper[4799]: I1129 05:46:59.051294 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac/rabbitmq/0.log" Nov 29 05:46:59 crc kubenswrapper[4799]: I1129 05:46:59.127053 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_0bfecf67-24fe-42fa-bad5-4a597ad9d3af/setup-container/0.log" Nov 29 05:46:59 crc kubenswrapper[4799]: I1129 05:46:59.350526 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_0bfecf67-24fe-42fa-bad5-4a597ad9d3af/setup-container/0.log" Nov 29 05:46:59 crc kubenswrapper[4799]: I1129 05:46:59.415117 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_0bfecf67-24fe-42fa-bad5-4a597ad9d3af/rabbitmq/0.log" Nov 29 05:46:59 crc kubenswrapper[4799]: I1129 05:46:59.418992 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-d68wr_0a726ce8-7d0d-4906-a2a4-e6f146dc4635/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 05:46:59 crc kubenswrapper[4799]: I1129 05:46:59.619412 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-vj9bq_3b36b813-5af5-4cc9-92c9-818aa2b99423/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 05:46:59 crc kubenswrapper[4799]: I1129 05:46:59.756802 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-h4p62_828374a6-f506-4e41-83fe-32bf0f6a4ed3/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 05:46:59 crc kubenswrapper[4799]: I1129 05:46:59.997093 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-9hpsr_3a84e2da-773c-4bf7-8f99-a84276791e78/ssh-known-hosts-edpm-deployment/0.log" Nov 29 05:47:00 crc kubenswrapper[4799]: I1129 05:47:00.047700 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_c4a30984-7c0c-493b-b752-b603232c837f/tempest-tests-tempest-tests-runner/0.log" Nov 29 05:47:00 crc kubenswrapper[4799]: I1129 05:47:00.238397 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_ec2916e5-c59e-4909-9c3c-da11e14e2413/test-operator-logs-container/0.log" Nov 29 05:47:00 crc kubenswrapper[4799]: I1129 05:47:00.295630 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-gnq8m_91c293ad-ab01-40e7-9fb6-b8ef99152e62/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 05:47:07 crc kubenswrapper[4799]: I1129 05:47:07.497740 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 05:47:07 crc kubenswrapper[4799]: I1129 05:47:07.498384 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 05:47:07 crc kubenswrapper[4799]: I1129 05:47:07.498445 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-26t88" Nov 29 05:47:07 crc kubenswrapper[4799]: I1129 05:47:07.499281 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bb3f916f31b5fc3b337d8f5f5faf98dbfdd362c77538109b1107019feab1d3dd"} pod="openshift-machine-config-operator/machine-config-daemon-26t88" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 29 05:47:07 crc kubenswrapper[4799]: I1129 05:47:07.499336 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" containerID="cri-o://bb3f916f31b5fc3b337d8f5f5faf98dbfdd362c77538109b1107019feab1d3dd" gracePeriod=600 Nov 29 05:47:07 crc kubenswrapper[4799]: I1129 05:47:07.912994 4799 generic.go:334] "Generic (PLEG): container finished" podID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerID="bb3f916f31b5fc3b337d8f5f5faf98dbfdd362c77538109b1107019feab1d3dd" exitCode=0 Nov 29 05:47:07 crc kubenswrapper[4799]: I1129 05:47:07.913048 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" event={"ID":"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8","Type":"ContainerDied","Data":"bb3f916f31b5fc3b337d8f5f5faf98dbfdd362c77538109b1107019feab1d3dd"} Nov 29 05:47:07 crc kubenswrapper[4799]: I1129 05:47:07.913376 4799 scope.go:117] "RemoveContainer" containerID="7baa15ad529e95d214463d6c72d4de622af57d5098e2ed42c89d7e00788ed77b" Nov 29 05:47:08 crc kubenswrapper[4799]: I1129 05:47:08.923001 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" event={"ID":"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8","Type":"ContainerStarted","Data":"b560e06b6de1edd9a7f603575cb604d0893a8a24eda3e04d2446b813f4e889f8"} Nov 29 05:47:16 crc kubenswrapper[4799]: I1129 05:47:16.632007 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_b6f5ef6d-bec8-4b82-8ebc-53952954134c/memcached/0.log" Nov 29 05:47:31 crc kubenswrapper[4799]: I1129 05:47:31.181046 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_232b41aac13e001dc857b9795d2b854afae448fcc1728af84c1ef8a484ctr76_e3b7bcee-1d1a-41a2-b1ea-221fe74998ef/util/0.log" Nov 29 05:47:31 crc kubenswrapper[4799]: I1129 05:47:31.299820 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_232b41aac13e001dc857b9795d2b854afae448fcc1728af84c1ef8a484ctr76_e3b7bcee-1d1a-41a2-b1ea-221fe74998ef/util/0.log" Nov 29 05:47:31 crc kubenswrapper[4799]: I1129 05:47:31.324765 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_232b41aac13e001dc857b9795d2b854afae448fcc1728af84c1ef8a484ctr76_e3b7bcee-1d1a-41a2-b1ea-221fe74998ef/pull/0.log" Nov 29 05:47:31 crc kubenswrapper[4799]: I1129 05:47:31.350116 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_232b41aac13e001dc857b9795d2b854afae448fcc1728af84c1ef8a484ctr76_e3b7bcee-1d1a-41a2-b1ea-221fe74998ef/pull/0.log" Nov 29 05:47:31 crc kubenswrapper[4799]: I1129 05:47:31.514958 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_232b41aac13e001dc857b9795d2b854afae448fcc1728af84c1ef8a484ctr76_e3b7bcee-1d1a-41a2-b1ea-221fe74998ef/pull/0.log" Nov 29 05:47:31 crc kubenswrapper[4799]: I1129 05:47:31.543447 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_232b41aac13e001dc857b9795d2b854afae448fcc1728af84c1ef8a484ctr76_e3b7bcee-1d1a-41a2-b1ea-221fe74998ef/util/0.log" Nov 29 05:47:31 crc kubenswrapper[4799]: I1129 05:47:31.574548 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_232b41aac13e001dc857b9795d2b854afae448fcc1728af84c1ef8a484ctr76_e3b7bcee-1d1a-41a2-b1ea-221fe74998ef/extract/0.log" Nov 29 05:47:31 crc kubenswrapper[4799]: I1129 05:47:31.700984 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-6vnhr_7232eadb-c887-440b-b6f9-bffa4a6dc871/kube-rbac-proxy/0.log" Nov 29 05:47:31 crc kubenswrapper[4799]: I1129 05:47:31.763492 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-7zddj_259abc1e-a30c-4a8b-acd1-18052862198e/kube-rbac-proxy/0.log" Nov 29 05:47:31 crc kubenswrapper[4799]: I1129 05:47:31.820547 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-6vnhr_7232eadb-c887-440b-b6f9-bffa4a6dc871/manager/0.log" Nov 29 05:47:31 crc kubenswrapper[4799]: I1129 05:47:31.943073 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-7zddj_259abc1e-a30c-4a8b-acd1-18052862198e/manager/0.log" Nov 29 05:47:32 crc kubenswrapper[4799]: I1129 05:47:32.022489 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-pkrhh_89038368-29a3-417e-8ffa-819e5901a52a/kube-rbac-proxy/0.log" Nov 29 05:47:32 crc kubenswrapper[4799]: I1129 05:47:32.087478 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-pkrhh_89038368-29a3-417e-8ffa-819e5901a52a/manager/0.log" Nov 29 05:47:32 crc kubenswrapper[4799]: I1129 05:47:32.188182 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-668d9c48b9-lkbfl_743ba96f-315f-4e0d-abb0-5a6389140c7b/kube-rbac-proxy/0.log" Nov 29 05:47:32 crc kubenswrapper[4799]: I1129 05:47:32.275511 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-668d9c48b9-lkbfl_743ba96f-315f-4e0d-abb0-5a6389140c7b/manager/0.log" Nov 29 05:47:32 crc kubenswrapper[4799]: I1129 05:47:32.295066 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-p555g_2798aeab-f9b9-4728-ad63-186b16afcc31/kube-rbac-proxy/0.log" Nov 29 05:47:32 crc kubenswrapper[4799]: I1129 05:47:32.401969 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-p555g_2798aeab-f9b9-4728-ad63-186b16afcc31/manager/0.log" Nov 29 05:47:32 crc kubenswrapper[4799]: I1129 05:47:32.487512 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-kxgw8_94ec7bc8-8413-4040-90dd-a5c9ce4ca98f/kube-rbac-proxy/0.log" Nov 29 05:47:32 crc kubenswrapper[4799]: I1129 05:47:32.557884 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-kxgw8_94ec7bc8-8413-4040-90dd-a5c9ce4ca98f/manager/0.log" Nov 29 05:47:32 crc kubenswrapper[4799]: I1129 05:47:32.642326 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-pn7wx_8c9ec062-c2b9-47a6-8d7a-eb9f00e63ef3/kube-rbac-proxy/0.log" Nov 29 05:47:32 crc kubenswrapper[4799]: I1129 05:47:32.805740 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-25fzb_ff72829c-dfd8-4ae2-b508-f89e40f654eb/kube-rbac-proxy/0.log" Nov 29 05:47:32 crc kubenswrapper[4799]: I1129 05:47:32.897285 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-25fzb_ff72829c-dfd8-4ae2-b508-f89e40f654eb/manager/0.log" Nov 29 05:47:32 crc kubenswrapper[4799]: I1129 05:47:32.904484 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-pn7wx_8c9ec062-c2b9-47a6-8d7a-eb9f00e63ef3/manager/0.log" Nov 29 05:47:33 crc kubenswrapper[4799]: I1129 05:47:33.017966 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-546d4bdf48-8p2z8_f696db58-bd41-4db0-9ec7-c4a4710e6b8e/kube-rbac-proxy/0.log" Nov 29 05:47:33 crc kubenswrapper[4799]: I1129 05:47:33.145669 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-84f754f96-t5gjl_4bf2a041-0f89-480e-88d2-ba630db189c1/kube-rbac-proxy/0.log" Nov 29 05:47:33 crc kubenswrapper[4799]: I1129 05:47:33.160656 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-546d4bdf48-8p2z8_f696db58-bd41-4db0-9ec7-c4a4710e6b8e/manager/0.log" Nov 29 05:47:33 crc kubenswrapper[4799]: I1129 05:47:33.290962 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-84f754f96-t5gjl_4bf2a041-0f89-480e-88d2-ba630db189c1/manager/0.log" Nov 29 05:47:33 crc kubenswrapper[4799]: I1129 05:47:33.340486 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-sfpmg_7851df67-88d5-45d8-804e-5d3bf3fdb3af/kube-rbac-proxy/0.log" Nov 29 05:47:33 crc kubenswrapper[4799]: I1129 05:47:33.373845 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-sfpmg_7851df67-88d5-45d8-804e-5d3bf3fdb3af/manager/0.log" Nov 29 05:47:33 crc kubenswrapper[4799]: I1129 05:47:33.547650 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-pc2zl_e1d949de-4ce7-4016-83fd-4c0ce003eb0f/kube-rbac-proxy/0.log" Nov 29 05:47:33 crc kubenswrapper[4799]: I1129 05:47:33.602878 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-pc2zl_e1d949de-4ce7-4016-83fd-4c0ce003eb0f/manager/0.log" Nov 29 05:47:33 crc kubenswrapper[4799]: I1129 05:47:33.711797 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-8nl87_bdcce387-cd82-490e-bd50-cf47158b1ea6/kube-rbac-proxy/0.log" Nov 29 05:47:33 crc kubenswrapper[4799]: I1129 05:47:33.808856 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-8nl87_bdcce387-cd82-490e-bd50-cf47158b1ea6/manager/0.log" Nov 29 05:47:33 crc kubenswrapper[4799]: I1129 05:47:33.851693 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-wstbs_bbe2e2fb-4619-405c-a9ad-4bdf727882cd/kube-rbac-proxy/0.log" Nov 29 05:47:33 crc kubenswrapper[4799]: I1129 05:47:33.908825 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-wstbs_bbe2e2fb-4619-405c-a9ad-4bdf727882cd/manager/0.log" Nov 29 05:47:34 crc kubenswrapper[4799]: I1129 05:47:34.074760 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4qnwnt_8e0677fe-7f53-4e50-881b-943fffe1c0ff/manager/0.log" Nov 29 05:47:34 crc kubenswrapper[4799]: I1129 05:47:34.081650 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4qnwnt_8e0677fe-7f53-4e50-881b-943fffe1c0ff/kube-rbac-proxy/0.log" Nov 29 05:47:34 crc kubenswrapper[4799]: I1129 05:47:34.380420 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-hr4n9_8a6a8fd6-7ba3-4457-9d23-b18414836f30/registry-server/0.log" Nov 29 05:47:34 crc kubenswrapper[4799]: I1129 05:47:34.474305 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-5bb85d56cf-bx84s_70a5864d-82f9-416d-a713-8e1bfa00733d/operator/0.log" Nov 29 05:47:34 crc kubenswrapper[4799]: I1129 05:47:34.574038 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-qfp99_888d5a2b-8a13-4042-8c43-59241cfdc088/kube-rbac-proxy/0.log" Nov 29 05:47:34 crc kubenswrapper[4799]: I1129 05:47:34.773243 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-qfp99_888d5a2b-8a13-4042-8c43-59241cfdc088/manager/0.log" Nov 29 05:47:34 crc kubenswrapper[4799]: I1129 05:47:34.788322 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-4p69t_861aea23-001f-4259-b69a-64b95db49c73/kube-rbac-proxy/0.log" Nov 29 05:47:34 crc kubenswrapper[4799]: I1129 05:47:34.835113 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-4p69t_861aea23-001f-4259-b69a-64b95db49c73/manager/0.log" Nov 29 05:47:34 crc kubenswrapper[4799]: I1129 05:47:34.982304 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-fsl4h_225a8e52-78ae-41eb-9c27-64edb57a135e/operator/0.log" Nov 29 05:47:35 crc kubenswrapper[4799]: I1129 05:47:35.063197 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-4xvn4_ac90f9ba-d2ef-44a1-84e7-8e809b5eef13/kube-rbac-proxy/0.log" Nov 29 05:47:35 crc kubenswrapper[4799]: I1129 05:47:35.212675 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-4xvn4_ac90f9ba-d2ef-44a1-84e7-8e809b5eef13/manager/0.log" Nov 29 05:47:35 crc kubenswrapper[4799]: I1129 05:47:35.295760 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-wv65n_06627027-315a-482d-9e26-32c5b4df115a/kube-rbac-proxy/0.log" Nov 29 05:47:35 crc kubenswrapper[4799]: I1129 05:47:35.467841 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-wv65n_06627027-315a-482d-9e26-32c5b4df115a/manager/0.log" Nov 29 05:47:35 crc kubenswrapper[4799]: I1129 05:47:35.529439 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-tbth4_1bb0507a-76c1-4a53-971c-dce960b71242/manager/0.log" Nov 29 05:47:35 crc kubenswrapper[4799]: I1129 05:47:35.538879 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-tbth4_1bb0507a-76c1-4a53-971c-dce960b71242/kube-rbac-proxy/0.log" Nov 29 05:47:35 crc kubenswrapper[4799]: I1129 05:47:35.601012 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-576b8587b8-cftqf_41eeb1eb-9b97-41fa-ad22-09d4d4b34abc/manager/0.log" Nov 29 05:47:35 crc kubenswrapper[4799]: I1129 05:47:35.712455 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-v4kv7_1edee957-6d50-48c2-bccc-e6702bcb5ce0/kube-rbac-proxy/0.log" Nov 29 05:47:35 crc kubenswrapper[4799]: I1129 05:47:35.740658 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-v4kv7_1edee957-6d50-48c2-bccc-e6702bcb5ce0/manager/0.log" Nov 29 05:47:53 crc kubenswrapper[4799]: I1129 05:47:53.919701 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-85llw_bfdd16e4-1cb8-4178-af49-4db764abc507/control-plane-machine-set-operator/0.log" Nov 29 05:47:54 crc kubenswrapper[4799]: I1129 05:47:54.119581 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-wznnv_daa3f105-2960-45a0-ab05-cb2cedea7e9f/kube-rbac-proxy/0.log" Nov 29 05:47:54 crc kubenswrapper[4799]: I1129 05:47:54.120152 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-wznnv_daa3f105-2960-45a0-ab05-cb2cedea7e9f/machine-api-operator/0.log" Nov 29 05:48:06 crc kubenswrapper[4799]: I1129 05:48:06.807696 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-68jsj_90513767-7b60-461e-bb69-727a8f25e15e/cert-manager-cainjector/0.log" Nov 29 05:48:06 crc kubenswrapper[4799]: I1129 05:48:06.834613 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-4hw5r_9347b23b-a5d2-40f4-8d55-a320f5a0c78a/cert-manager-controller/0.log" Nov 29 05:48:07 crc kubenswrapper[4799]: I1129 05:48:07.016428 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-b9lmh_28341ac5-a271-4435-a855-512a548beb96/cert-manager-webhook/0.log" Nov 29 05:48:19 crc kubenswrapper[4799]: I1129 05:48:19.547500 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-j5cvx_bb0e8eeb-0667-40e9-b814-33a66c0e78a0/nmstate-console-plugin/0.log" Nov 29 05:48:19 crc kubenswrapper[4799]: I1129 05:48:19.755938 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-7k5v8_767c1d25-4a7d-49b3-8180-8f46793649be/nmstate-handler/0.log" Nov 29 05:48:19 crc kubenswrapper[4799]: I1129 05:48:19.861366 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-2955x_e96ac85b-718a-45dc-be96-b514bffa1431/nmstate-metrics/0.log" Nov 29 05:48:19 crc kubenswrapper[4799]: I1129 05:48:19.862952 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-2955x_e96ac85b-718a-45dc-be96-b514bffa1431/kube-rbac-proxy/0.log" Nov 29 05:48:20 crc kubenswrapper[4799]: I1129 05:48:20.047341 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-7wvt6_9924770d-6e30-4da4-959a-40b3eade5bf3/nmstate-operator/0.log" Nov 29 05:48:20 crc kubenswrapper[4799]: I1129 05:48:20.065324 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-2kxrf_337961b2-38e7-4e32-831a-0c970535fecd/nmstate-webhook/0.log" Nov 29 05:48:34 crc kubenswrapper[4799]: I1129 05:48:34.138135 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-27dxl_9b8b8dfd-90f2-44b4-921a-e56563c76655/kube-rbac-proxy/0.log" Nov 29 05:48:34 crc kubenswrapper[4799]: I1129 05:48:34.285236 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-27dxl_9b8b8dfd-90f2-44b4-921a-e56563c76655/controller/0.log" Nov 29 05:48:34 crc kubenswrapper[4799]: I1129 05:48:34.369356 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95pk_bae7578a-afd9-462d-b824-beae428d9edf/cp-frr-files/0.log" Nov 29 05:48:34 crc kubenswrapper[4799]: I1129 05:48:34.549734 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95pk_bae7578a-afd9-462d-b824-beae428d9edf/cp-metrics/0.log" Nov 29 05:48:34 crc kubenswrapper[4799]: I1129 05:48:34.558374 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95pk_bae7578a-afd9-462d-b824-beae428d9edf/cp-reloader/0.log" Nov 29 05:48:34 crc kubenswrapper[4799]: I1129 05:48:34.619604 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95pk_bae7578a-afd9-462d-b824-beae428d9edf/cp-frr-files/0.log" Nov 29 05:48:34 crc kubenswrapper[4799]: I1129 05:48:34.630237 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95pk_bae7578a-afd9-462d-b824-beae428d9edf/cp-reloader/0.log" Nov 29 05:48:34 crc kubenswrapper[4799]: I1129 05:48:34.764171 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95pk_bae7578a-afd9-462d-b824-beae428d9edf/cp-frr-files/0.log" Nov 29 05:48:34 crc kubenswrapper[4799]: I1129 05:48:34.772526 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95pk_bae7578a-afd9-462d-b824-beae428d9edf/cp-metrics/0.log" Nov 29 05:48:34 crc kubenswrapper[4799]: I1129 05:48:34.814073 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95pk_bae7578a-afd9-462d-b824-beae428d9edf/cp-reloader/0.log" Nov 29 05:48:34 crc kubenswrapper[4799]: I1129 05:48:34.817546 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95pk_bae7578a-afd9-462d-b824-beae428d9edf/cp-metrics/0.log" Nov 29 05:48:35 crc kubenswrapper[4799]: I1129 05:48:35.378359 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95pk_bae7578a-afd9-462d-b824-beae428d9edf/cp-reloader/0.log" Nov 29 05:48:35 crc kubenswrapper[4799]: I1129 05:48:35.402586 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95pk_bae7578a-afd9-462d-b824-beae428d9edf/cp-frr-files/0.log" Nov 29 05:48:35 crc kubenswrapper[4799]: I1129 05:48:35.423327 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95pk_bae7578a-afd9-462d-b824-beae428d9edf/cp-metrics/0.log" Nov 29 05:48:35 crc kubenswrapper[4799]: I1129 05:48:35.425414 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95pk_bae7578a-afd9-462d-b824-beae428d9edf/controller/0.log" Nov 29 05:48:35 crc kubenswrapper[4799]: I1129 05:48:35.568751 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95pk_bae7578a-afd9-462d-b824-beae428d9edf/frr-metrics/0.log" Nov 29 05:48:35 crc kubenswrapper[4799]: I1129 05:48:35.650339 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95pk_bae7578a-afd9-462d-b824-beae428d9edf/kube-rbac-proxy/0.log" Nov 29 05:48:35 crc kubenswrapper[4799]: I1129 05:48:35.669731 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95pk_bae7578a-afd9-462d-b824-beae428d9edf/kube-rbac-proxy-frr/0.log" Nov 29 05:48:35 crc kubenswrapper[4799]: I1129 05:48:35.845481 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95pk_bae7578a-afd9-462d-b824-beae428d9edf/reloader/0.log" Nov 29 05:48:35 crc kubenswrapper[4799]: I1129 05:48:35.853576 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-j9pvr_95a0cc49-7fb2-4ed4-85a5-50d4248f07aa/frr-k8s-webhook-server/0.log" Nov 29 05:48:36 crc kubenswrapper[4799]: I1129 05:48:36.217630 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-6887d9747b-t425t_cabb2055-e31c-4a75-a9f9-d403684efeec/manager/0.log" Nov 29 05:48:36 crc kubenswrapper[4799]: I1129 05:48:36.362957 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-78b9fb955c-gptcp_dd389a73-2aa5-4f69-8ff1-5a3e7a421c1e/webhook-server/0.log" Nov 29 05:48:36 crc kubenswrapper[4799]: I1129 05:48:36.418768 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-tc7fc_0845b467-e744-408a-b2e7-d61b027724dd/kube-rbac-proxy/0.log" Nov 29 05:48:37 crc kubenswrapper[4799]: I1129 05:48:37.239430 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95pk_bae7578a-afd9-462d-b824-beae428d9edf/frr/0.log" Nov 29 05:48:37 crc kubenswrapper[4799]: I1129 05:48:37.359861 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-tc7fc_0845b467-e744-408a-b2e7-d61b027724dd/speaker/0.log" Nov 29 05:48:48 crc kubenswrapper[4799]: I1129 05:48:48.630769 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftbmdt_01a9ddbf-bda8-4f39-b22e-400f1a1505b0/util/0.log" Nov 29 05:48:48 crc kubenswrapper[4799]: I1129 05:48:48.864622 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftbmdt_01a9ddbf-bda8-4f39-b22e-400f1a1505b0/util/0.log" Nov 29 05:48:48 crc kubenswrapper[4799]: I1129 05:48:48.885631 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftbmdt_01a9ddbf-bda8-4f39-b22e-400f1a1505b0/pull/0.log" Nov 29 05:48:48 crc kubenswrapper[4799]: I1129 05:48:48.951910 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftbmdt_01a9ddbf-bda8-4f39-b22e-400f1a1505b0/pull/0.log" Nov 29 05:48:49 crc kubenswrapper[4799]: I1129 05:48:49.081529 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftbmdt_01a9ddbf-bda8-4f39-b22e-400f1a1505b0/extract/0.log" Nov 29 05:48:49 crc kubenswrapper[4799]: I1129 05:48:49.088307 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftbmdt_01a9ddbf-bda8-4f39-b22e-400f1a1505b0/pull/0.log" Nov 29 05:48:49 crc kubenswrapper[4799]: I1129 05:48:49.091883 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftbmdt_01a9ddbf-bda8-4f39-b22e-400f1a1505b0/util/0.log" Nov 29 05:48:49 crc kubenswrapper[4799]: I1129 05:48:49.234106 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83flptb_c94962e4-e51f-4e30-9c0e-75a19dfa16d9/util/0.log" Nov 29 05:48:49 crc kubenswrapper[4799]: I1129 05:48:49.412568 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83flptb_c94962e4-e51f-4e30-9c0e-75a19dfa16d9/util/0.log" Nov 29 05:48:49 crc kubenswrapper[4799]: I1129 05:48:49.438106 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83flptb_c94962e4-e51f-4e30-9c0e-75a19dfa16d9/pull/0.log" Nov 29 05:48:49 crc kubenswrapper[4799]: I1129 05:48:49.460615 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83flptb_c94962e4-e51f-4e30-9c0e-75a19dfa16d9/pull/0.log" Nov 29 05:48:49 crc kubenswrapper[4799]: I1129 05:48:49.636415 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83flptb_c94962e4-e51f-4e30-9c0e-75a19dfa16d9/pull/0.log" Nov 29 05:48:49 crc kubenswrapper[4799]: I1129 05:48:49.638641 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83flptb_c94962e4-e51f-4e30-9c0e-75a19dfa16d9/util/0.log" Nov 29 05:48:49 crc kubenswrapper[4799]: I1129 05:48:49.685694 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83flptb_c94962e4-e51f-4e30-9c0e-75a19dfa16d9/extract/0.log" Nov 29 05:48:49 crc kubenswrapper[4799]: I1129 05:48:49.796918 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pmw7d_cfabf29e-ad4c-4ded-b754-5b02a25ff34b/extract-utilities/0.log" Nov 29 05:48:49 crc kubenswrapper[4799]: I1129 05:48:49.990958 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pmw7d_cfabf29e-ad4c-4ded-b754-5b02a25ff34b/extract-content/0.log" Nov 29 05:48:49 crc kubenswrapper[4799]: I1129 05:48:49.996654 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pmw7d_cfabf29e-ad4c-4ded-b754-5b02a25ff34b/extract-utilities/0.log" Nov 29 05:48:50 crc kubenswrapper[4799]: I1129 05:48:50.000225 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pmw7d_cfabf29e-ad4c-4ded-b754-5b02a25ff34b/extract-content/0.log" Nov 29 05:48:50 crc kubenswrapper[4799]: I1129 05:48:50.199375 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pmw7d_cfabf29e-ad4c-4ded-b754-5b02a25ff34b/extract-content/0.log" Nov 29 05:48:50 crc kubenswrapper[4799]: I1129 05:48:50.241641 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pmw7d_cfabf29e-ad4c-4ded-b754-5b02a25ff34b/extract-utilities/0.log" Nov 29 05:48:50 crc kubenswrapper[4799]: I1129 05:48:50.432990 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-wc7hk_3642ba6f-6029-462d-9441-903056f9886a/extract-utilities/0.log" Nov 29 05:48:50 crc kubenswrapper[4799]: I1129 05:48:50.657210 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-wc7hk_3642ba6f-6029-462d-9441-903056f9886a/extract-content/0.log" Nov 29 05:48:50 crc kubenswrapper[4799]: I1129 05:48:50.686192 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-wc7hk_3642ba6f-6029-462d-9441-903056f9886a/extract-content/0.log" Nov 29 05:48:50 crc kubenswrapper[4799]: I1129 05:48:50.698105 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-wc7hk_3642ba6f-6029-462d-9441-903056f9886a/extract-utilities/0.log" Nov 29 05:48:50 crc kubenswrapper[4799]: I1129 05:48:50.770877 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pmw7d_cfabf29e-ad4c-4ded-b754-5b02a25ff34b/registry-server/0.log" Nov 29 05:48:50 crc kubenswrapper[4799]: I1129 05:48:50.848218 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-wc7hk_3642ba6f-6029-462d-9441-903056f9886a/extract-utilities/0.log" Nov 29 05:48:50 crc kubenswrapper[4799]: I1129 05:48:50.882599 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-wc7hk_3642ba6f-6029-462d-9441-903056f9886a/extract-content/0.log" Nov 29 05:48:50 crc kubenswrapper[4799]: I1129 05:48:50.992551 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-wc7hk_3642ba6f-6029-462d-9441-903056f9886a/registry-server/0.log" Nov 29 05:48:51 crc kubenswrapper[4799]: I1129 05:48:51.064052 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-szfkm_72063360-3635-4677-8d2a-8d639af3ad6d/marketplace-operator/0.log" Nov 29 05:48:51 crc kubenswrapper[4799]: I1129 05:48:51.160995 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qwxcg_6ee05926-5af3-4fac-9772-8e1484c4de17/extract-utilities/0.log" Nov 29 05:48:51 crc kubenswrapper[4799]: I1129 05:48:51.314259 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qwxcg_6ee05926-5af3-4fac-9772-8e1484c4de17/extract-utilities/0.log" Nov 29 05:48:51 crc kubenswrapper[4799]: I1129 05:48:51.331019 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qwxcg_6ee05926-5af3-4fac-9772-8e1484c4de17/extract-content/0.log" Nov 29 05:48:51 crc kubenswrapper[4799]: I1129 05:48:51.354109 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qwxcg_6ee05926-5af3-4fac-9772-8e1484c4de17/extract-content/0.log" Nov 29 05:48:51 crc kubenswrapper[4799]: I1129 05:48:51.480459 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qwxcg_6ee05926-5af3-4fac-9772-8e1484c4de17/extract-content/0.log" Nov 29 05:48:51 crc kubenswrapper[4799]: I1129 05:48:51.499526 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qwxcg_6ee05926-5af3-4fac-9772-8e1484c4de17/extract-utilities/0.log" Nov 29 05:48:51 crc kubenswrapper[4799]: I1129 05:48:51.617038 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qwxcg_6ee05926-5af3-4fac-9772-8e1484c4de17/registry-server/0.log" Nov 29 05:48:51 crc kubenswrapper[4799]: I1129 05:48:51.671347 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xngqr_47f0c357-1fc0-4646-ae33-b4497ac972e7/extract-utilities/0.log" Nov 29 05:48:51 crc kubenswrapper[4799]: I1129 05:48:51.911229 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xngqr_47f0c357-1fc0-4646-ae33-b4497ac972e7/extract-content/0.log" Nov 29 05:48:51 crc kubenswrapper[4799]: I1129 05:48:51.913489 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xngqr_47f0c357-1fc0-4646-ae33-b4497ac972e7/extract-utilities/0.log" Nov 29 05:48:51 crc kubenswrapper[4799]: I1129 05:48:51.931348 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xngqr_47f0c357-1fc0-4646-ae33-b4497ac972e7/extract-content/0.log" Nov 29 05:48:52 crc kubenswrapper[4799]: I1129 05:48:52.048655 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xngqr_47f0c357-1fc0-4646-ae33-b4497ac972e7/extract-content/0.log" Nov 29 05:48:52 crc kubenswrapper[4799]: I1129 05:48:52.067744 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xngqr_47f0c357-1fc0-4646-ae33-b4497ac972e7/extract-utilities/0.log" Nov 29 05:48:52 crc kubenswrapper[4799]: I1129 05:48:52.573087 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xngqr_47f0c357-1fc0-4646-ae33-b4497ac972e7/registry-server/0.log" Nov 29 05:49:37 crc kubenswrapper[4799]: I1129 05:49:37.497350 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 05:49:37 crc kubenswrapper[4799]: I1129 05:49:37.498024 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 05:50:07 crc kubenswrapper[4799]: I1129 05:50:07.497207 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 05:50:07 crc kubenswrapper[4799]: I1129 05:50:07.497774 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 05:50:32 crc kubenswrapper[4799]: I1129 05:50:32.569986 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hp7vg"] Nov 29 05:50:32 crc kubenswrapper[4799]: E1129 05:50:32.573955 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="147b56fb-2ad7-4807-b27b-de550b69cfda" containerName="container-00" Nov 29 05:50:32 crc kubenswrapper[4799]: I1129 05:50:32.574008 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="147b56fb-2ad7-4807-b27b-de550b69cfda" containerName="container-00" Nov 29 05:50:32 crc kubenswrapper[4799]: I1129 05:50:32.574379 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="147b56fb-2ad7-4807-b27b-de550b69cfda" containerName="container-00" Nov 29 05:50:32 crc kubenswrapper[4799]: I1129 05:50:32.578287 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hp7vg" Nov 29 05:50:32 crc kubenswrapper[4799]: I1129 05:50:32.584486 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hp7vg"] Nov 29 05:50:32 crc kubenswrapper[4799]: I1129 05:50:32.586364 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/955e0b59-ac77-433b-a2df-a170e0da483d-catalog-content\") pod \"redhat-marketplace-hp7vg\" (UID: \"955e0b59-ac77-433b-a2df-a170e0da483d\") " pod="openshift-marketplace/redhat-marketplace-hp7vg" Nov 29 05:50:32 crc kubenswrapper[4799]: I1129 05:50:32.586539 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnw7r\" (UniqueName: \"kubernetes.io/projected/955e0b59-ac77-433b-a2df-a170e0da483d-kube-api-access-xnw7r\") pod \"redhat-marketplace-hp7vg\" (UID: \"955e0b59-ac77-433b-a2df-a170e0da483d\") " pod="openshift-marketplace/redhat-marketplace-hp7vg" Nov 29 05:50:32 crc kubenswrapper[4799]: I1129 05:50:32.586628 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/955e0b59-ac77-433b-a2df-a170e0da483d-utilities\") pod \"redhat-marketplace-hp7vg\" (UID: \"955e0b59-ac77-433b-a2df-a170e0da483d\") " pod="openshift-marketplace/redhat-marketplace-hp7vg" Nov 29 05:50:32 crc kubenswrapper[4799]: I1129 05:50:32.689043 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/955e0b59-ac77-433b-a2df-a170e0da483d-catalog-content\") pod \"redhat-marketplace-hp7vg\" (UID: \"955e0b59-ac77-433b-a2df-a170e0da483d\") " pod="openshift-marketplace/redhat-marketplace-hp7vg" Nov 29 05:50:32 crc kubenswrapper[4799]: I1129 05:50:32.690005 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnw7r\" (UniqueName: \"kubernetes.io/projected/955e0b59-ac77-433b-a2df-a170e0da483d-kube-api-access-xnw7r\") pod \"redhat-marketplace-hp7vg\" (UID: \"955e0b59-ac77-433b-a2df-a170e0da483d\") " pod="openshift-marketplace/redhat-marketplace-hp7vg" Nov 29 05:50:32 crc kubenswrapper[4799]: I1129 05:50:32.690075 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/955e0b59-ac77-433b-a2df-a170e0da483d-utilities\") pod \"redhat-marketplace-hp7vg\" (UID: \"955e0b59-ac77-433b-a2df-a170e0da483d\") " pod="openshift-marketplace/redhat-marketplace-hp7vg" Nov 29 05:50:32 crc kubenswrapper[4799]: I1129 05:50:32.689783 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/955e0b59-ac77-433b-a2df-a170e0da483d-catalog-content\") pod \"redhat-marketplace-hp7vg\" (UID: \"955e0b59-ac77-433b-a2df-a170e0da483d\") " pod="openshift-marketplace/redhat-marketplace-hp7vg" Nov 29 05:50:32 crc kubenswrapper[4799]: I1129 05:50:32.695402 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/955e0b59-ac77-433b-a2df-a170e0da483d-utilities\") pod \"redhat-marketplace-hp7vg\" (UID: \"955e0b59-ac77-433b-a2df-a170e0da483d\") " pod="openshift-marketplace/redhat-marketplace-hp7vg" Nov 29 05:50:32 crc kubenswrapper[4799]: I1129 05:50:32.711773 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnw7r\" (UniqueName: \"kubernetes.io/projected/955e0b59-ac77-433b-a2df-a170e0da483d-kube-api-access-xnw7r\") pod \"redhat-marketplace-hp7vg\" (UID: \"955e0b59-ac77-433b-a2df-a170e0da483d\") " pod="openshift-marketplace/redhat-marketplace-hp7vg" Nov 29 05:50:32 crc kubenswrapper[4799]: I1129 05:50:32.911999 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hp7vg" Nov 29 05:50:33 crc kubenswrapper[4799]: I1129 05:50:33.386592 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hp7vg"] Nov 29 05:50:33 crc kubenswrapper[4799]: I1129 05:50:33.750870 4799 generic.go:334] "Generic (PLEG): container finished" podID="955e0b59-ac77-433b-a2df-a170e0da483d" containerID="de4c994333e568ef0ffa3b58d8931f641c179ed87ad41e1120afead53ca1414a" exitCode=0 Nov 29 05:50:33 crc kubenswrapper[4799]: I1129 05:50:33.750925 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hp7vg" event={"ID":"955e0b59-ac77-433b-a2df-a170e0da483d","Type":"ContainerDied","Data":"de4c994333e568ef0ffa3b58d8931f641c179ed87ad41e1120afead53ca1414a"} Nov 29 05:50:33 crc kubenswrapper[4799]: I1129 05:50:33.750979 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hp7vg" event={"ID":"955e0b59-ac77-433b-a2df-a170e0da483d","Type":"ContainerStarted","Data":"3381be6b57e1008c8342102daba219602bcb21249fd83976005d2d31959e7b9f"} Nov 29 05:50:33 crc kubenswrapper[4799]: I1129 05:50:33.752763 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 29 05:50:35 crc kubenswrapper[4799]: I1129 05:50:35.769603 4799 generic.go:334] "Generic (PLEG): container finished" podID="955e0b59-ac77-433b-a2df-a170e0da483d" containerID="07ef20084d4d2b1cbb027daf433c81110b8993aa3ef6145d5aa1a556a650915c" exitCode=0 Nov 29 05:50:35 crc kubenswrapper[4799]: I1129 05:50:35.769664 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hp7vg" event={"ID":"955e0b59-ac77-433b-a2df-a170e0da483d","Type":"ContainerDied","Data":"07ef20084d4d2b1cbb027daf433c81110b8993aa3ef6145d5aa1a556a650915c"} Nov 29 05:50:36 crc kubenswrapper[4799]: I1129 05:50:36.782939 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hp7vg" event={"ID":"955e0b59-ac77-433b-a2df-a170e0da483d","Type":"ContainerStarted","Data":"abfb9ad1be1b09890366d3c7164441ebaf120b1660fe5ae6921177b72ba116bd"} Nov 29 05:50:36 crc kubenswrapper[4799]: I1129 05:50:36.804607 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hp7vg" podStartSLOduration=2.233532784 podStartE2EDuration="4.804586577s" podCreationTimestamp="2025-11-29 05:50:32 +0000 UTC" firstStartedPulling="2025-11-29 05:50:33.752475152 +0000 UTC m=+4309.395405552" lastFinishedPulling="2025-11-29 05:50:36.323528945 +0000 UTC m=+4311.966459345" observedRunningTime="2025-11-29 05:50:36.800459369 +0000 UTC m=+4312.443389769" watchObservedRunningTime="2025-11-29 05:50:36.804586577 +0000 UTC m=+4312.447516977" Nov 29 05:50:37 crc kubenswrapper[4799]: I1129 05:50:37.498049 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 05:50:37 crc kubenswrapper[4799]: I1129 05:50:37.498108 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 05:50:37 crc kubenswrapper[4799]: I1129 05:50:37.498150 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-26t88" Nov 29 05:50:37 crc kubenswrapper[4799]: I1129 05:50:37.498823 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b560e06b6de1edd9a7f603575cb604d0893a8a24eda3e04d2446b813f4e889f8"} pod="openshift-machine-config-operator/machine-config-daemon-26t88" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 29 05:50:37 crc kubenswrapper[4799]: I1129 05:50:37.498881 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" containerID="cri-o://b560e06b6de1edd9a7f603575cb604d0893a8a24eda3e04d2446b813f4e889f8" gracePeriod=600 Nov 29 05:50:37 crc kubenswrapper[4799]: E1129 05:50:37.620330 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:50:37 crc kubenswrapper[4799]: I1129 05:50:37.806557 4799 generic.go:334] "Generic (PLEG): container finished" podID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerID="b560e06b6de1edd9a7f603575cb604d0893a8a24eda3e04d2446b813f4e889f8" exitCode=0 Nov 29 05:50:37 crc kubenswrapper[4799]: I1129 05:50:37.806630 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" event={"ID":"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8","Type":"ContainerDied","Data":"b560e06b6de1edd9a7f603575cb604d0893a8a24eda3e04d2446b813f4e889f8"} Nov 29 05:50:37 crc kubenswrapper[4799]: I1129 05:50:37.806686 4799 scope.go:117] "RemoveContainer" containerID="bb3f916f31b5fc3b337d8f5f5faf98dbfdd362c77538109b1107019feab1d3dd" Nov 29 05:50:37 crc kubenswrapper[4799]: I1129 05:50:37.807390 4799 scope.go:117] "RemoveContainer" containerID="b560e06b6de1edd9a7f603575cb604d0893a8a24eda3e04d2446b813f4e889f8" Nov 29 05:50:37 crc kubenswrapper[4799]: E1129 05:50:37.807697 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:50:42 crc kubenswrapper[4799]: I1129 05:50:42.914735 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hp7vg" Nov 29 05:50:42 crc kubenswrapper[4799]: I1129 05:50:42.915350 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hp7vg" Nov 29 05:50:42 crc kubenswrapper[4799]: I1129 05:50:42.972343 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hp7vg" Nov 29 05:50:43 crc kubenswrapper[4799]: I1129 05:50:43.908042 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hp7vg" Nov 29 05:50:43 crc kubenswrapper[4799]: I1129 05:50:43.960058 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hp7vg"] Nov 29 05:50:45 crc kubenswrapper[4799]: I1129 05:50:45.875883 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hp7vg" podUID="955e0b59-ac77-433b-a2df-a170e0da483d" containerName="registry-server" containerID="cri-o://abfb9ad1be1b09890366d3c7164441ebaf120b1660fe5ae6921177b72ba116bd" gracePeriod=2 Nov 29 05:50:46 crc kubenswrapper[4799]: I1129 05:50:46.383638 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hp7vg" Nov 29 05:50:46 crc kubenswrapper[4799]: I1129 05:50:46.571578 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/955e0b59-ac77-433b-a2df-a170e0da483d-catalog-content\") pod \"955e0b59-ac77-433b-a2df-a170e0da483d\" (UID: \"955e0b59-ac77-433b-a2df-a170e0da483d\") " Nov 29 05:50:46 crc kubenswrapper[4799]: I1129 05:50:46.571973 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xnw7r\" (UniqueName: \"kubernetes.io/projected/955e0b59-ac77-433b-a2df-a170e0da483d-kube-api-access-xnw7r\") pod \"955e0b59-ac77-433b-a2df-a170e0da483d\" (UID: \"955e0b59-ac77-433b-a2df-a170e0da483d\") " Nov 29 05:50:46 crc kubenswrapper[4799]: I1129 05:50:46.572001 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/955e0b59-ac77-433b-a2df-a170e0da483d-utilities\") pod \"955e0b59-ac77-433b-a2df-a170e0da483d\" (UID: \"955e0b59-ac77-433b-a2df-a170e0da483d\") " Nov 29 05:50:46 crc kubenswrapper[4799]: I1129 05:50:46.573685 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/955e0b59-ac77-433b-a2df-a170e0da483d-utilities" (OuterVolumeSpecName: "utilities") pod "955e0b59-ac77-433b-a2df-a170e0da483d" (UID: "955e0b59-ac77-433b-a2df-a170e0da483d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:50:46 crc kubenswrapper[4799]: I1129 05:50:46.585693 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/955e0b59-ac77-433b-a2df-a170e0da483d-kube-api-access-xnw7r" (OuterVolumeSpecName: "kube-api-access-xnw7r") pod "955e0b59-ac77-433b-a2df-a170e0da483d" (UID: "955e0b59-ac77-433b-a2df-a170e0da483d"). InnerVolumeSpecName "kube-api-access-xnw7r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:50:46 crc kubenswrapper[4799]: I1129 05:50:46.609139 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/955e0b59-ac77-433b-a2df-a170e0da483d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "955e0b59-ac77-433b-a2df-a170e0da483d" (UID: "955e0b59-ac77-433b-a2df-a170e0da483d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:50:46 crc kubenswrapper[4799]: I1129 05:50:46.675130 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xnw7r\" (UniqueName: \"kubernetes.io/projected/955e0b59-ac77-433b-a2df-a170e0da483d-kube-api-access-xnw7r\") on node \"crc\" DevicePath \"\"" Nov 29 05:50:46 crc kubenswrapper[4799]: I1129 05:50:46.675178 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/955e0b59-ac77-433b-a2df-a170e0da483d-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 05:50:46 crc kubenswrapper[4799]: I1129 05:50:46.675189 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/955e0b59-ac77-433b-a2df-a170e0da483d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 05:50:46 crc kubenswrapper[4799]: E1129 05:50:46.839925 4799 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod955e0b59_ac77_433b_a2df_a170e0da483d.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod955e0b59_ac77_433b_a2df_a170e0da483d.slice/crio-3381be6b57e1008c8342102daba219602bcb21249fd83976005d2d31959e7b9f\": RecentStats: unable to find data in memory cache]" Nov 29 05:50:46 crc kubenswrapper[4799]: I1129 05:50:46.884632 4799 generic.go:334] "Generic (PLEG): container finished" podID="955e0b59-ac77-433b-a2df-a170e0da483d" containerID="abfb9ad1be1b09890366d3c7164441ebaf120b1660fe5ae6921177b72ba116bd" exitCode=0 Nov 29 05:50:46 crc kubenswrapper[4799]: I1129 05:50:46.884668 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hp7vg" Nov 29 05:50:46 crc kubenswrapper[4799]: I1129 05:50:46.884675 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hp7vg" event={"ID":"955e0b59-ac77-433b-a2df-a170e0da483d","Type":"ContainerDied","Data":"abfb9ad1be1b09890366d3c7164441ebaf120b1660fe5ae6921177b72ba116bd"} Nov 29 05:50:46 crc kubenswrapper[4799]: I1129 05:50:46.884723 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hp7vg" event={"ID":"955e0b59-ac77-433b-a2df-a170e0da483d","Type":"ContainerDied","Data":"3381be6b57e1008c8342102daba219602bcb21249fd83976005d2d31959e7b9f"} Nov 29 05:50:46 crc kubenswrapper[4799]: I1129 05:50:46.884743 4799 scope.go:117] "RemoveContainer" containerID="abfb9ad1be1b09890366d3c7164441ebaf120b1660fe5ae6921177b72ba116bd" Nov 29 05:50:46 crc kubenswrapper[4799]: I1129 05:50:46.888749 4799 generic.go:334] "Generic (PLEG): container finished" podID="3fac1ef9-0b5b-49e4-a5df-c95b3982a363" containerID="23b13520f503adbd1e25255e588391fa59ac2ebe583d8fd278eebc1a65099261" exitCode=0 Nov 29 05:50:46 crc kubenswrapper[4799]: I1129 05:50:46.888815 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ll7z5/must-gather-p24k7" event={"ID":"3fac1ef9-0b5b-49e4-a5df-c95b3982a363","Type":"ContainerDied","Data":"23b13520f503adbd1e25255e588391fa59ac2ebe583d8fd278eebc1a65099261"} Nov 29 05:50:46 crc kubenswrapper[4799]: I1129 05:50:46.890350 4799 scope.go:117] "RemoveContainer" containerID="23b13520f503adbd1e25255e588391fa59ac2ebe583d8fd278eebc1a65099261" Nov 29 05:50:46 crc kubenswrapper[4799]: I1129 05:50:46.911563 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hp7vg"] Nov 29 05:50:46 crc kubenswrapper[4799]: I1129 05:50:46.918218 4799 scope.go:117] "RemoveContainer" containerID="07ef20084d4d2b1cbb027daf433c81110b8993aa3ef6145d5aa1a556a650915c" Nov 29 05:50:46 crc kubenswrapper[4799]: I1129 05:50:46.923243 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hp7vg"] Nov 29 05:50:46 crc kubenswrapper[4799]: I1129 05:50:46.951417 4799 scope.go:117] "RemoveContainer" containerID="de4c994333e568ef0ffa3b58d8931f641c179ed87ad41e1120afead53ca1414a" Nov 29 05:50:46 crc kubenswrapper[4799]: I1129 05:50:46.985233 4799 scope.go:117] "RemoveContainer" containerID="abfb9ad1be1b09890366d3c7164441ebaf120b1660fe5ae6921177b72ba116bd" Nov 29 05:50:46 crc kubenswrapper[4799]: E1129 05:50:46.985980 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"abfb9ad1be1b09890366d3c7164441ebaf120b1660fe5ae6921177b72ba116bd\": container with ID starting with abfb9ad1be1b09890366d3c7164441ebaf120b1660fe5ae6921177b72ba116bd not found: ID does not exist" containerID="abfb9ad1be1b09890366d3c7164441ebaf120b1660fe5ae6921177b72ba116bd" Nov 29 05:50:46 crc kubenswrapper[4799]: I1129 05:50:46.986016 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abfb9ad1be1b09890366d3c7164441ebaf120b1660fe5ae6921177b72ba116bd"} err="failed to get container status \"abfb9ad1be1b09890366d3c7164441ebaf120b1660fe5ae6921177b72ba116bd\": rpc error: code = NotFound desc = could not find container \"abfb9ad1be1b09890366d3c7164441ebaf120b1660fe5ae6921177b72ba116bd\": container with ID starting with abfb9ad1be1b09890366d3c7164441ebaf120b1660fe5ae6921177b72ba116bd not found: ID does not exist" Nov 29 05:50:46 crc kubenswrapper[4799]: I1129 05:50:46.986042 4799 scope.go:117] "RemoveContainer" containerID="07ef20084d4d2b1cbb027daf433c81110b8993aa3ef6145d5aa1a556a650915c" Nov 29 05:50:46 crc kubenswrapper[4799]: E1129 05:50:46.986536 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07ef20084d4d2b1cbb027daf433c81110b8993aa3ef6145d5aa1a556a650915c\": container with ID starting with 07ef20084d4d2b1cbb027daf433c81110b8993aa3ef6145d5aa1a556a650915c not found: ID does not exist" containerID="07ef20084d4d2b1cbb027daf433c81110b8993aa3ef6145d5aa1a556a650915c" Nov 29 05:50:46 crc kubenswrapper[4799]: I1129 05:50:46.986568 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07ef20084d4d2b1cbb027daf433c81110b8993aa3ef6145d5aa1a556a650915c"} err="failed to get container status \"07ef20084d4d2b1cbb027daf433c81110b8993aa3ef6145d5aa1a556a650915c\": rpc error: code = NotFound desc = could not find container \"07ef20084d4d2b1cbb027daf433c81110b8993aa3ef6145d5aa1a556a650915c\": container with ID starting with 07ef20084d4d2b1cbb027daf433c81110b8993aa3ef6145d5aa1a556a650915c not found: ID does not exist" Nov 29 05:50:46 crc kubenswrapper[4799]: I1129 05:50:46.986585 4799 scope.go:117] "RemoveContainer" containerID="de4c994333e568ef0ffa3b58d8931f641c179ed87ad41e1120afead53ca1414a" Nov 29 05:50:46 crc kubenswrapper[4799]: E1129 05:50:46.986869 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de4c994333e568ef0ffa3b58d8931f641c179ed87ad41e1120afead53ca1414a\": container with ID starting with de4c994333e568ef0ffa3b58d8931f641c179ed87ad41e1120afead53ca1414a not found: ID does not exist" containerID="de4c994333e568ef0ffa3b58d8931f641c179ed87ad41e1120afead53ca1414a" Nov 29 05:50:46 crc kubenswrapper[4799]: I1129 05:50:46.986898 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de4c994333e568ef0ffa3b58d8931f641c179ed87ad41e1120afead53ca1414a"} err="failed to get container status \"de4c994333e568ef0ffa3b58d8931f641c179ed87ad41e1120afead53ca1414a\": rpc error: code = NotFound desc = could not find container \"de4c994333e568ef0ffa3b58d8931f641c179ed87ad41e1120afead53ca1414a\": container with ID starting with de4c994333e568ef0ffa3b58d8931f641c179ed87ad41e1120afead53ca1414a not found: ID does not exist" Nov 29 05:50:47 crc kubenswrapper[4799]: I1129 05:50:47.136246 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ll7z5_must-gather-p24k7_3fac1ef9-0b5b-49e4-a5df-c95b3982a363/gather/0.log" Nov 29 05:50:48 crc kubenswrapper[4799]: I1129 05:50:48.673271 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="955e0b59-ac77-433b-a2df-a170e0da483d" path="/var/lib/kubelet/pods/955e0b59-ac77-433b-a2df-a170e0da483d/volumes" Nov 29 05:50:49 crc kubenswrapper[4799]: E1129 05:50:49.283630 4799 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.83:33986->38.102.83.83:33865: write tcp 38.102.83.83:33986->38.102.83.83:33865: write: broken pipe Nov 29 05:50:50 crc kubenswrapper[4799]: I1129 05:50:50.659604 4799 scope.go:117] "RemoveContainer" containerID="b560e06b6de1edd9a7f603575cb604d0893a8a24eda3e04d2446b813f4e889f8" Nov 29 05:50:50 crc kubenswrapper[4799]: E1129 05:50:50.660333 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:50:54 crc kubenswrapper[4799]: I1129 05:50:54.770067 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-ll7z5/must-gather-p24k7"] Nov 29 05:50:54 crc kubenswrapper[4799]: I1129 05:50:54.770872 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-ll7z5/must-gather-p24k7" podUID="3fac1ef9-0b5b-49e4-a5df-c95b3982a363" containerName="copy" containerID="cri-o://817aa9869e57608dd06023a2defac38f44f2f7ac794773a4e2f4739ed2efc421" gracePeriod=2 Nov 29 05:50:54 crc kubenswrapper[4799]: I1129 05:50:54.780097 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-ll7z5/must-gather-p24k7"] Nov 29 05:50:54 crc kubenswrapper[4799]: I1129 05:50:54.965449 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ll7z5_must-gather-p24k7_3fac1ef9-0b5b-49e4-a5df-c95b3982a363/copy/0.log" Nov 29 05:50:54 crc kubenswrapper[4799]: I1129 05:50:54.966111 4799 generic.go:334] "Generic (PLEG): container finished" podID="3fac1ef9-0b5b-49e4-a5df-c95b3982a363" containerID="817aa9869e57608dd06023a2defac38f44f2f7ac794773a4e2f4739ed2efc421" exitCode=143 Nov 29 05:50:55 crc kubenswrapper[4799]: I1129 05:50:55.165350 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ll7z5_must-gather-p24k7_3fac1ef9-0b5b-49e4-a5df-c95b3982a363/copy/0.log" Nov 29 05:50:55 crc kubenswrapper[4799]: I1129 05:50:55.165806 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ll7z5/must-gather-p24k7" Nov 29 05:50:55 crc kubenswrapper[4799]: I1129 05:50:55.357641 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3fac1ef9-0b5b-49e4-a5df-c95b3982a363-must-gather-output\") pod \"3fac1ef9-0b5b-49e4-a5df-c95b3982a363\" (UID: \"3fac1ef9-0b5b-49e4-a5df-c95b3982a363\") " Nov 29 05:50:55 crc kubenswrapper[4799]: I1129 05:50:55.357769 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lt7hr\" (UniqueName: \"kubernetes.io/projected/3fac1ef9-0b5b-49e4-a5df-c95b3982a363-kube-api-access-lt7hr\") pod \"3fac1ef9-0b5b-49e4-a5df-c95b3982a363\" (UID: \"3fac1ef9-0b5b-49e4-a5df-c95b3982a363\") " Nov 29 05:50:55 crc kubenswrapper[4799]: I1129 05:50:55.364308 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fac1ef9-0b5b-49e4-a5df-c95b3982a363-kube-api-access-lt7hr" (OuterVolumeSpecName: "kube-api-access-lt7hr") pod "3fac1ef9-0b5b-49e4-a5df-c95b3982a363" (UID: "3fac1ef9-0b5b-49e4-a5df-c95b3982a363"). InnerVolumeSpecName "kube-api-access-lt7hr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:50:55 crc kubenswrapper[4799]: I1129 05:50:55.460303 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lt7hr\" (UniqueName: \"kubernetes.io/projected/3fac1ef9-0b5b-49e4-a5df-c95b3982a363-kube-api-access-lt7hr\") on node \"crc\" DevicePath \"\"" Nov 29 05:50:55 crc kubenswrapper[4799]: I1129 05:50:55.516389 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3fac1ef9-0b5b-49e4-a5df-c95b3982a363-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "3fac1ef9-0b5b-49e4-a5df-c95b3982a363" (UID: "3fac1ef9-0b5b-49e4-a5df-c95b3982a363"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:50:55 crc kubenswrapper[4799]: I1129 05:50:55.561980 4799 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3fac1ef9-0b5b-49e4-a5df-c95b3982a363-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 29 05:50:55 crc kubenswrapper[4799]: I1129 05:50:55.977477 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ll7z5_must-gather-p24k7_3fac1ef9-0b5b-49e4-a5df-c95b3982a363/copy/0.log" Nov 29 05:50:55 crc kubenswrapper[4799]: I1129 05:50:55.977994 4799 scope.go:117] "RemoveContainer" containerID="817aa9869e57608dd06023a2defac38f44f2f7ac794773a4e2f4739ed2efc421" Nov 29 05:50:55 crc kubenswrapper[4799]: I1129 05:50:55.978068 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ll7z5/must-gather-p24k7" Nov 29 05:50:56 crc kubenswrapper[4799]: I1129 05:50:56.007148 4799 scope.go:117] "RemoveContainer" containerID="23b13520f503adbd1e25255e588391fa59ac2ebe583d8fd278eebc1a65099261" Nov 29 05:50:56 crc kubenswrapper[4799]: I1129 05:50:56.670567 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3fac1ef9-0b5b-49e4-a5df-c95b3982a363" path="/var/lib/kubelet/pods/3fac1ef9-0b5b-49e4-a5df-c95b3982a363/volumes" Nov 29 05:51:01 crc kubenswrapper[4799]: I1129 05:51:01.659267 4799 scope.go:117] "RemoveContainer" containerID="b560e06b6de1edd9a7f603575cb604d0893a8a24eda3e04d2446b813f4e889f8" Nov 29 05:51:01 crc kubenswrapper[4799]: E1129 05:51:01.660330 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:51:15 crc kubenswrapper[4799]: I1129 05:51:15.659044 4799 scope.go:117] "RemoveContainer" containerID="b560e06b6de1edd9a7f603575cb604d0893a8a24eda3e04d2446b813f4e889f8" Nov 29 05:51:15 crc kubenswrapper[4799]: E1129 05:51:15.659883 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:51:28 crc kubenswrapper[4799]: I1129 05:51:28.658826 4799 scope.go:117] "RemoveContainer" containerID="b560e06b6de1edd9a7f603575cb604d0893a8a24eda3e04d2446b813f4e889f8" Nov 29 05:51:28 crc kubenswrapper[4799]: E1129 05:51:28.659642 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:51:42 crc kubenswrapper[4799]: I1129 05:51:42.659215 4799 scope.go:117] "RemoveContainer" containerID="b560e06b6de1edd9a7f603575cb604d0893a8a24eda3e04d2446b813f4e889f8" Nov 29 05:51:42 crc kubenswrapper[4799]: E1129 05:51:42.660738 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:51:53 crc kubenswrapper[4799]: I1129 05:51:53.658827 4799 scope.go:117] "RemoveContainer" containerID="b560e06b6de1edd9a7f603575cb604d0893a8a24eda3e04d2446b813f4e889f8" Nov 29 05:51:53 crc kubenswrapper[4799]: E1129 05:51:53.659470 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:52:04 crc kubenswrapper[4799]: I1129 05:52:04.665347 4799 scope.go:117] "RemoveContainer" containerID="b560e06b6de1edd9a7f603575cb604d0893a8a24eda3e04d2446b813f4e889f8" Nov 29 05:52:04 crc kubenswrapper[4799]: E1129 05:52:04.666388 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:52:18 crc kubenswrapper[4799]: I1129 05:52:18.659319 4799 scope.go:117] "RemoveContainer" containerID="b560e06b6de1edd9a7f603575cb604d0893a8a24eda3e04d2446b813f4e889f8" Nov 29 05:52:18 crc kubenswrapper[4799]: E1129 05:52:18.660110 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:52:32 crc kubenswrapper[4799]: I1129 05:52:32.659583 4799 scope.go:117] "RemoveContainer" containerID="b560e06b6de1edd9a7f603575cb604d0893a8a24eda3e04d2446b813f4e889f8" Nov 29 05:52:32 crc kubenswrapper[4799]: E1129 05:52:32.660622 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:52:45 crc kubenswrapper[4799]: I1129 05:52:45.659432 4799 scope.go:117] "RemoveContainer" containerID="b560e06b6de1edd9a7f603575cb604d0893a8a24eda3e04d2446b813f4e889f8" Nov 29 05:52:45 crc kubenswrapper[4799]: E1129 05:52:45.660423 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:52:59 crc kubenswrapper[4799]: I1129 05:52:59.404622 4799 scope.go:117] "RemoveContainer" containerID="f7453ef5077138b52ed9b6487e7eb3d2cc030232cabb278103bcf9d493387c34" Nov 29 05:52:59 crc kubenswrapper[4799]: I1129 05:52:59.659188 4799 scope.go:117] "RemoveContainer" containerID="b560e06b6de1edd9a7f603575cb604d0893a8a24eda3e04d2446b813f4e889f8" Nov 29 05:52:59 crc kubenswrapper[4799]: E1129 05:52:59.659503 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:53:11 crc kubenswrapper[4799]: I1129 05:53:11.659222 4799 scope.go:117] "RemoveContainer" containerID="b560e06b6de1edd9a7f603575cb604d0893a8a24eda3e04d2446b813f4e889f8" Nov 29 05:53:11 crc kubenswrapper[4799]: E1129 05:53:11.659911 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:53:25 crc kubenswrapper[4799]: I1129 05:53:25.659047 4799 scope.go:117] "RemoveContainer" containerID="b560e06b6de1edd9a7f603575cb604d0893a8a24eda3e04d2446b813f4e889f8" Nov 29 05:53:25 crc kubenswrapper[4799]: E1129 05:53:25.659820 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:53:39 crc kubenswrapper[4799]: I1129 05:53:39.659520 4799 scope.go:117] "RemoveContainer" containerID="b560e06b6de1edd9a7f603575cb604d0893a8a24eda3e04d2446b813f4e889f8" Nov 29 05:53:39 crc kubenswrapper[4799]: E1129 05:53:39.660255 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:53:46 crc kubenswrapper[4799]: I1129 05:53:46.960913 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dxhxp"] Nov 29 05:53:46 crc kubenswrapper[4799]: E1129 05:53:46.962030 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fac1ef9-0b5b-49e4-a5df-c95b3982a363" containerName="gather" Nov 29 05:53:46 crc kubenswrapper[4799]: I1129 05:53:46.962047 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fac1ef9-0b5b-49e4-a5df-c95b3982a363" containerName="gather" Nov 29 05:53:46 crc kubenswrapper[4799]: E1129 05:53:46.962078 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="955e0b59-ac77-433b-a2df-a170e0da483d" containerName="registry-server" Nov 29 05:53:46 crc kubenswrapper[4799]: I1129 05:53:46.962086 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="955e0b59-ac77-433b-a2df-a170e0da483d" containerName="registry-server" Nov 29 05:53:46 crc kubenswrapper[4799]: E1129 05:53:46.962121 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="955e0b59-ac77-433b-a2df-a170e0da483d" containerName="extract-utilities" Nov 29 05:53:46 crc kubenswrapper[4799]: I1129 05:53:46.962129 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="955e0b59-ac77-433b-a2df-a170e0da483d" containerName="extract-utilities" Nov 29 05:53:46 crc kubenswrapper[4799]: E1129 05:53:46.962144 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fac1ef9-0b5b-49e4-a5df-c95b3982a363" containerName="copy" Nov 29 05:53:46 crc kubenswrapper[4799]: I1129 05:53:46.962151 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fac1ef9-0b5b-49e4-a5df-c95b3982a363" containerName="copy" Nov 29 05:53:46 crc kubenswrapper[4799]: E1129 05:53:46.962166 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="955e0b59-ac77-433b-a2df-a170e0da483d" containerName="extract-content" Nov 29 05:53:46 crc kubenswrapper[4799]: I1129 05:53:46.962174 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="955e0b59-ac77-433b-a2df-a170e0da483d" containerName="extract-content" Nov 29 05:53:46 crc kubenswrapper[4799]: I1129 05:53:46.962420 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fac1ef9-0b5b-49e4-a5df-c95b3982a363" containerName="gather" Nov 29 05:53:46 crc kubenswrapper[4799]: I1129 05:53:46.962445 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fac1ef9-0b5b-49e4-a5df-c95b3982a363" containerName="copy" Nov 29 05:53:46 crc kubenswrapper[4799]: I1129 05:53:46.962462 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="955e0b59-ac77-433b-a2df-a170e0da483d" containerName="registry-server" Nov 29 05:53:46 crc kubenswrapper[4799]: I1129 05:53:46.964476 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dxhxp" Nov 29 05:53:46 crc kubenswrapper[4799]: I1129 05:53:46.979238 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dxhxp"] Nov 29 05:53:46 crc kubenswrapper[4799]: I1129 05:53:46.985652 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb075ee4-ae05-449f-abac-14ea3f614f2f-utilities\") pod \"redhat-operators-dxhxp\" (UID: \"cb075ee4-ae05-449f-abac-14ea3f614f2f\") " pod="openshift-marketplace/redhat-operators-dxhxp" Nov 29 05:53:46 crc kubenswrapper[4799]: I1129 05:53:46.988183 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h972k\" (UniqueName: \"kubernetes.io/projected/cb075ee4-ae05-449f-abac-14ea3f614f2f-kube-api-access-h972k\") pod \"redhat-operators-dxhxp\" (UID: \"cb075ee4-ae05-449f-abac-14ea3f614f2f\") " pod="openshift-marketplace/redhat-operators-dxhxp" Nov 29 05:53:46 crc kubenswrapper[4799]: I1129 05:53:46.988305 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb075ee4-ae05-449f-abac-14ea3f614f2f-catalog-content\") pod \"redhat-operators-dxhxp\" (UID: \"cb075ee4-ae05-449f-abac-14ea3f614f2f\") " pod="openshift-marketplace/redhat-operators-dxhxp" Nov 29 05:53:47 crc kubenswrapper[4799]: I1129 05:53:47.090337 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb075ee4-ae05-449f-abac-14ea3f614f2f-utilities\") pod \"redhat-operators-dxhxp\" (UID: \"cb075ee4-ae05-449f-abac-14ea3f614f2f\") " pod="openshift-marketplace/redhat-operators-dxhxp" Nov 29 05:53:47 crc kubenswrapper[4799]: I1129 05:53:47.090652 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h972k\" (UniqueName: \"kubernetes.io/projected/cb075ee4-ae05-449f-abac-14ea3f614f2f-kube-api-access-h972k\") pod \"redhat-operators-dxhxp\" (UID: \"cb075ee4-ae05-449f-abac-14ea3f614f2f\") " pod="openshift-marketplace/redhat-operators-dxhxp" Nov 29 05:53:47 crc kubenswrapper[4799]: I1129 05:53:47.090957 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb075ee4-ae05-449f-abac-14ea3f614f2f-utilities\") pod \"redhat-operators-dxhxp\" (UID: \"cb075ee4-ae05-449f-abac-14ea3f614f2f\") " pod="openshift-marketplace/redhat-operators-dxhxp" Nov 29 05:53:47 crc kubenswrapper[4799]: I1129 05:53:47.091052 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb075ee4-ae05-449f-abac-14ea3f614f2f-catalog-content\") pod \"redhat-operators-dxhxp\" (UID: \"cb075ee4-ae05-449f-abac-14ea3f614f2f\") " pod="openshift-marketplace/redhat-operators-dxhxp" Nov 29 05:53:47 crc kubenswrapper[4799]: I1129 05:53:47.091396 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb075ee4-ae05-449f-abac-14ea3f614f2f-catalog-content\") pod \"redhat-operators-dxhxp\" (UID: \"cb075ee4-ae05-449f-abac-14ea3f614f2f\") " pod="openshift-marketplace/redhat-operators-dxhxp" Nov 29 05:53:47 crc kubenswrapper[4799]: I1129 05:53:47.114482 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h972k\" (UniqueName: \"kubernetes.io/projected/cb075ee4-ae05-449f-abac-14ea3f614f2f-kube-api-access-h972k\") pod \"redhat-operators-dxhxp\" (UID: \"cb075ee4-ae05-449f-abac-14ea3f614f2f\") " pod="openshift-marketplace/redhat-operators-dxhxp" Nov 29 05:53:47 crc kubenswrapper[4799]: I1129 05:53:47.296839 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dxhxp" Nov 29 05:53:47 crc kubenswrapper[4799]: I1129 05:53:47.760280 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dxhxp"] Nov 29 05:53:48 crc kubenswrapper[4799]: I1129 05:53:48.645715 4799 generic.go:334] "Generic (PLEG): container finished" podID="cb075ee4-ae05-449f-abac-14ea3f614f2f" containerID="aff747c64c96600960e92fdf0e7308b360330f95c51345ef48c133fd570d1f27" exitCode=0 Nov 29 05:53:48 crc kubenswrapper[4799]: I1129 05:53:48.645773 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dxhxp" event={"ID":"cb075ee4-ae05-449f-abac-14ea3f614f2f","Type":"ContainerDied","Data":"aff747c64c96600960e92fdf0e7308b360330f95c51345ef48c133fd570d1f27"} Nov 29 05:53:48 crc kubenswrapper[4799]: I1129 05:53:48.646287 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dxhxp" event={"ID":"cb075ee4-ae05-449f-abac-14ea3f614f2f","Type":"ContainerStarted","Data":"1449c52087017efb82822daabee4bf233a96853c4d851986c0a2db1863088174"} Nov 29 05:53:49 crc kubenswrapper[4799]: I1129 05:53:49.657465 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dxhxp" event={"ID":"cb075ee4-ae05-449f-abac-14ea3f614f2f","Type":"ContainerStarted","Data":"397b0bdba6fb759747e2d89ca48e2b4030df290b8216add6e905ad116c1c2ebb"} Nov 29 05:53:49 crc kubenswrapper[4799]: I1129 05:53:49.835297 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-t9ms2/must-gather-c5k8l"] Nov 29 05:53:49 crc kubenswrapper[4799]: I1129 05:53:49.837346 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t9ms2/must-gather-c5k8l" Nov 29 05:53:49 crc kubenswrapper[4799]: I1129 05:53:49.844145 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-t9ms2"/"kube-root-ca.crt" Nov 29 05:53:49 crc kubenswrapper[4799]: I1129 05:53:49.844365 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-t9ms2"/"default-dockercfg-w6n8p" Nov 29 05:53:49 crc kubenswrapper[4799]: I1129 05:53:49.845563 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-t9ms2"/"openshift-service-ca.crt" Nov 29 05:53:49 crc kubenswrapper[4799]: I1129 05:53:49.869541 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-t9ms2/must-gather-c5k8l"] Nov 29 05:53:49 crc kubenswrapper[4799]: I1129 05:53:49.956189 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/426fd24d-c567-40bf-902b-ce637dbf9f84-must-gather-output\") pod \"must-gather-c5k8l\" (UID: \"426fd24d-c567-40bf-902b-ce637dbf9f84\") " pod="openshift-must-gather-t9ms2/must-gather-c5k8l" Nov 29 05:53:49 crc kubenswrapper[4799]: I1129 05:53:49.956333 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbgxg\" (UniqueName: \"kubernetes.io/projected/426fd24d-c567-40bf-902b-ce637dbf9f84-kube-api-access-wbgxg\") pod \"must-gather-c5k8l\" (UID: \"426fd24d-c567-40bf-902b-ce637dbf9f84\") " pod="openshift-must-gather-t9ms2/must-gather-c5k8l" Nov 29 05:53:50 crc kubenswrapper[4799]: I1129 05:53:50.057953 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/426fd24d-c567-40bf-902b-ce637dbf9f84-must-gather-output\") pod \"must-gather-c5k8l\" (UID: \"426fd24d-c567-40bf-902b-ce637dbf9f84\") " pod="openshift-must-gather-t9ms2/must-gather-c5k8l" Nov 29 05:53:50 crc kubenswrapper[4799]: I1129 05:53:50.058104 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbgxg\" (UniqueName: \"kubernetes.io/projected/426fd24d-c567-40bf-902b-ce637dbf9f84-kube-api-access-wbgxg\") pod \"must-gather-c5k8l\" (UID: \"426fd24d-c567-40bf-902b-ce637dbf9f84\") " pod="openshift-must-gather-t9ms2/must-gather-c5k8l" Nov 29 05:53:50 crc kubenswrapper[4799]: I1129 05:53:50.058416 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/426fd24d-c567-40bf-902b-ce637dbf9f84-must-gather-output\") pod \"must-gather-c5k8l\" (UID: \"426fd24d-c567-40bf-902b-ce637dbf9f84\") " pod="openshift-must-gather-t9ms2/must-gather-c5k8l" Nov 29 05:53:50 crc kubenswrapper[4799]: I1129 05:53:50.081660 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbgxg\" (UniqueName: \"kubernetes.io/projected/426fd24d-c567-40bf-902b-ce637dbf9f84-kube-api-access-wbgxg\") pod \"must-gather-c5k8l\" (UID: \"426fd24d-c567-40bf-902b-ce637dbf9f84\") " pod="openshift-must-gather-t9ms2/must-gather-c5k8l" Nov 29 05:53:50 crc kubenswrapper[4799]: I1129 05:53:50.160034 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t9ms2/must-gather-c5k8l" Nov 29 05:53:50 crc kubenswrapper[4799]: I1129 05:53:50.669974 4799 generic.go:334] "Generic (PLEG): container finished" podID="cb075ee4-ae05-449f-abac-14ea3f614f2f" containerID="397b0bdba6fb759747e2d89ca48e2b4030df290b8216add6e905ad116c1c2ebb" exitCode=0 Nov 29 05:53:50 crc kubenswrapper[4799]: I1129 05:53:50.670071 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dxhxp" event={"ID":"cb075ee4-ae05-449f-abac-14ea3f614f2f","Type":"ContainerDied","Data":"397b0bdba6fb759747e2d89ca48e2b4030df290b8216add6e905ad116c1c2ebb"} Nov 29 05:53:50 crc kubenswrapper[4799]: W1129 05:53:50.702867 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod426fd24d_c567_40bf_902b_ce637dbf9f84.slice/crio-5fcdc328661aaf367c17ef9a47b23c7965b10bb029012235f989502b0128332e WatchSource:0}: Error finding container 5fcdc328661aaf367c17ef9a47b23c7965b10bb029012235f989502b0128332e: Status 404 returned error can't find the container with id 5fcdc328661aaf367c17ef9a47b23c7965b10bb029012235f989502b0128332e Nov 29 05:53:50 crc kubenswrapper[4799]: I1129 05:53:50.718554 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-t9ms2/must-gather-c5k8l"] Nov 29 05:53:51 crc kubenswrapper[4799]: I1129 05:53:51.678598 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t9ms2/must-gather-c5k8l" event={"ID":"426fd24d-c567-40bf-902b-ce637dbf9f84","Type":"ContainerStarted","Data":"50442bc8a06a2575e37ee856944ca8d1b43b4b863b1de800c9168c4f5fbf45d4"} Nov 29 05:53:51 crc kubenswrapper[4799]: I1129 05:53:51.679044 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t9ms2/must-gather-c5k8l" event={"ID":"426fd24d-c567-40bf-902b-ce637dbf9f84","Type":"ContainerStarted","Data":"c690a584291d73690cd800625a57f4a66b22b00d44a4b37de8494177dfc7e871"} Nov 29 05:53:51 crc kubenswrapper[4799]: I1129 05:53:51.679056 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t9ms2/must-gather-c5k8l" event={"ID":"426fd24d-c567-40bf-902b-ce637dbf9f84","Type":"ContainerStarted","Data":"5fcdc328661aaf367c17ef9a47b23c7965b10bb029012235f989502b0128332e"} Nov 29 05:53:51 crc kubenswrapper[4799]: I1129 05:53:51.682084 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dxhxp" event={"ID":"cb075ee4-ae05-449f-abac-14ea3f614f2f","Type":"ContainerStarted","Data":"6641b638b7c7bef08c558fce13fd39a5dfe050050b44603676245206a56ffc00"} Nov 29 05:53:51 crc kubenswrapper[4799]: I1129 05:53:51.725706 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-t9ms2/must-gather-c5k8l" podStartSLOduration=2.725689318 podStartE2EDuration="2.725689318s" podCreationTimestamp="2025-11-29 05:53:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 05:53:51.704366577 +0000 UTC m=+4507.347296977" watchObservedRunningTime="2025-11-29 05:53:51.725689318 +0000 UTC m=+4507.368619718" Nov 29 05:53:51 crc kubenswrapper[4799]: I1129 05:53:51.726220 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dxhxp" podStartSLOduration=3.26265718 podStartE2EDuration="5.72621199s" podCreationTimestamp="2025-11-29 05:53:46 +0000 UTC" firstStartedPulling="2025-11-29 05:53:48.647516549 +0000 UTC m=+4504.290446949" lastFinishedPulling="2025-11-29 05:53:51.111071359 +0000 UTC m=+4506.754001759" observedRunningTime="2025-11-29 05:53:51.720584256 +0000 UTC m=+4507.363514656" watchObservedRunningTime="2025-11-29 05:53:51.72621199 +0000 UTC m=+4507.369142390" Nov 29 05:53:53 crc kubenswrapper[4799]: I1129 05:53:53.659852 4799 scope.go:117] "RemoveContainer" containerID="b560e06b6de1edd9a7f603575cb604d0893a8a24eda3e04d2446b813f4e889f8" Nov 29 05:53:53 crc kubenswrapper[4799]: E1129 05:53:53.662716 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:53:54 crc kubenswrapper[4799]: E1129 05:53:54.348445 4799 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.83:45836->38.102.83.83:33865: write tcp 38.102.83.83:45836->38.102.83.83:33865: write: connection reset by peer Nov 29 05:53:55 crc kubenswrapper[4799]: I1129 05:53:55.979032 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-t9ms2/crc-debug-dndqf"] Nov 29 05:53:55 crc kubenswrapper[4799]: I1129 05:53:55.980679 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t9ms2/crc-debug-dndqf" Nov 29 05:53:56 crc kubenswrapper[4799]: I1129 05:53:56.093713 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/474c86be-9e78-48eb-b5b0-60db74e967e3-host\") pod \"crc-debug-dndqf\" (UID: \"474c86be-9e78-48eb-b5b0-60db74e967e3\") " pod="openshift-must-gather-t9ms2/crc-debug-dndqf" Nov 29 05:53:56 crc kubenswrapper[4799]: I1129 05:53:56.094389 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrpx5\" (UniqueName: \"kubernetes.io/projected/474c86be-9e78-48eb-b5b0-60db74e967e3-kube-api-access-jrpx5\") pod \"crc-debug-dndqf\" (UID: \"474c86be-9e78-48eb-b5b0-60db74e967e3\") " pod="openshift-must-gather-t9ms2/crc-debug-dndqf" Nov 29 05:53:56 crc kubenswrapper[4799]: I1129 05:53:56.197527 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/474c86be-9e78-48eb-b5b0-60db74e967e3-host\") pod \"crc-debug-dndqf\" (UID: \"474c86be-9e78-48eb-b5b0-60db74e967e3\") " pod="openshift-must-gather-t9ms2/crc-debug-dndqf" Nov 29 05:53:56 crc kubenswrapper[4799]: I1129 05:53:56.197642 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/474c86be-9e78-48eb-b5b0-60db74e967e3-host\") pod \"crc-debug-dndqf\" (UID: \"474c86be-9e78-48eb-b5b0-60db74e967e3\") " pod="openshift-must-gather-t9ms2/crc-debug-dndqf" Nov 29 05:53:56 crc kubenswrapper[4799]: I1129 05:53:56.197646 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrpx5\" (UniqueName: \"kubernetes.io/projected/474c86be-9e78-48eb-b5b0-60db74e967e3-kube-api-access-jrpx5\") pod \"crc-debug-dndqf\" (UID: \"474c86be-9e78-48eb-b5b0-60db74e967e3\") " pod="openshift-must-gather-t9ms2/crc-debug-dndqf" Nov 29 05:53:56 crc kubenswrapper[4799]: I1129 05:53:56.218427 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrpx5\" (UniqueName: \"kubernetes.io/projected/474c86be-9e78-48eb-b5b0-60db74e967e3-kube-api-access-jrpx5\") pod \"crc-debug-dndqf\" (UID: \"474c86be-9e78-48eb-b5b0-60db74e967e3\") " pod="openshift-must-gather-t9ms2/crc-debug-dndqf" Nov 29 05:53:56 crc kubenswrapper[4799]: I1129 05:53:56.303052 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t9ms2/crc-debug-dndqf" Nov 29 05:53:56 crc kubenswrapper[4799]: I1129 05:53:56.740987 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t9ms2/crc-debug-dndqf" event={"ID":"474c86be-9e78-48eb-b5b0-60db74e967e3","Type":"ContainerStarted","Data":"ef0b744ea4f6579b61788bfaef631d414b3f2d29987941ed66f0fbe9df9ad447"} Nov 29 05:53:57 crc kubenswrapper[4799]: I1129 05:53:57.297810 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dxhxp" Nov 29 05:53:57 crc kubenswrapper[4799]: I1129 05:53:57.298128 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dxhxp" Nov 29 05:53:57 crc kubenswrapper[4799]: I1129 05:53:57.618829 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dxhxp" Nov 29 05:53:57 crc kubenswrapper[4799]: I1129 05:53:57.751118 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t9ms2/crc-debug-dndqf" event={"ID":"474c86be-9e78-48eb-b5b0-60db74e967e3","Type":"ContainerStarted","Data":"499d0333bd30be0734900a58a7daf348badb369ce4c1cb2598b375bd380a1b12"} Nov 29 05:53:57 crc kubenswrapper[4799]: I1129 05:53:57.768836 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-t9ms2/crc-debug-dndqf" podStartSLOduration=2.768812736 podStartE2EDuration="2.768812736s" podCreationTimestamp="2025-11-29 05:53:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 05:53:57.763829326 +0000 UTC m=+4513.406759726" watchObservedRunningTime="2025-11-29 05:53:57.768812736 +0000 UTC m=+4513.411743146" Nov 29 05:53:57 crc kubenswrapper[4799]: I1129 05:53:57.812492 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dxhxp" Nov 29 05:53:57 crc kubenswrapper[4799]: I1129 05:53:57.867100 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dxhxp"] Nov 29 05:53:59 crc kubenswrapper[4799]: I1129 05:53:59.765270 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dxhxp" podUID="cb075ee4-ae05-449f-abac-14ea3f614f2f" containerName="registry-server" containerID="cri-o://6641b638b7c7bef08c558fce13fd39a5dfe050050b44603676245206a56ffc00" gracePeriod=2 Nov 29 05:54:00 crc kubenswrapper[4799]: I1129 05:54:00.283874 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6qxrx"] Nov 29 05:54:00 crc kubenswrapper[4799]: I1129 05:54:00.291403 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6qxrx" Nov 29 05:54:00 crc kubenswrapper[4799]: I1129 05:54:00.297194 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1fc3fae-d6bc-4f40-b01c-06eac982f674-catalog-content\") pod \"certified-operators-6qxrx\" (UID: \"e1fc3fae-d6bc-4f40-b01c-06eac982f674\") " pod="openshift-marketplace/certified-operators-6qxrx" Nov 29 05:54:00 crc kubenswrapper[4799]: I1129 05:54:00.297522 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6c6n\" (UniqueName: \"kubernetes.io/projected/e1fc3fae-d6bc-4f40-b01c-06eac982f674-kube-api-access-t6c6n\") pod \"certified-operators-6qxrx\" (UID: \"e1fc3fae-d6bc-4f40-b01c-06eac982f674\") " pod="openshift-marketplace/certified-operators-6qxrx" Nov 29 05:54:00 crc kubenswrapper[4799]: I1129 05:54:00.297892 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1fc3fae-d6bc-4f40-b01c-06eac982f674-utilities\") pod \"certified-operators-6qxrx\" (UID: \"e1fc3fae-d6bc-4f40-b01c-06eac982f674\") " pod="openshift-marketplace/certified-operators-6qxrx" Nov 29 05:54:00 crc kubenswrapper[4799]: I1129 05:54:00.325703 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6qxrx"] Nov 29 05:54:00 crc kubenswrapper[4799]: I1129 05:54:00.348417 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dxhxp" Nov 29 05:54:00 crc kubenswrapper[4799]: I1129 05:54:00.401276 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb075ee4-ae05-449f-abac-14ea3f614f2f-utilities\") pod \"cb075ee4-ae05-449f-abac-14ea3f614f2f\" (UID: \"cb075ee4-ae05-449f-abac-14ea3f614f2f\") " Nov 29 05:54:00 crc kubenswrapper[4799]: I1129 05:54:00.402697 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb075ee4-ae05-449f-abac-14ea3f614f2f-catalog-content\") pod \"cb075ee4-ae05-449f-abac-14ea3f614f2f\" (UID: \"cb075ee4-ae05-449f-abac-14ea3f614f2f\") " Nov 29 05:54:00 crc kubenswrapper[4799]: I1129 05:54:00.402871 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h972k\" (UniqueName: \"kubernetes.io/projected/cb075ee4-ae05-449f-abac-14ea3f614f2f-kube-api-access-h972k\") pod \"cb075ee4-ae05-449f-abac-14ea3f614f2f\" (UID: \"cb075ee4-ae05-449f-abac-14ea3f614f2f\") " Nov 29 05:54:00 crc kubenswrapper[4799]: I1129 05:54:00.404155 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb075ee4-ae05-449f-abac-14ea3f614f2f-utilities" (OuterVolumeSpecName: "utilities") pod "cb075ee4-ae05-449f-abac-14ea3f614f2f" (UID: "cb075ee4-ae05-449f-abac-14ea3f614f2f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:54:00 crc kubenswrapper[4799]: I1129 05:54:00.404499 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1fc3fae-d6bc-4f40-b01c-06eac982f674-utilities\") pod \"certified-operators-6qxrx\" (UID: \"e1fc3fae-d6bc-4f40-b01c-06eac982f674\") " pod="openshift-marketplace/certified-operators-6qxrx" Nov 29 05:54:00 crc kubenswrapper[4799]: I1129 05:54:00.404690 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1fc3fae-d6bc-4f40-b01c-06eac982f674-catalog-content\") pod \"certified-operators-6qxrx\" (UID: \"e1fc3fae-d6bc-4f40-b01c-06eac982f674\") " pod="openshift-marketplace/certified-operators-6qxrx" Nov 29 05:54:00 crc kubenswrapper[4799]: I1129 05:54:00.404877 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6c6n\" (UniqueName: \"kubernetes.io/projected/e1fc3fae-d6bc-4f40-b01c-06eac982f674-kube-api-access-t6c6n\") pod \"certified-operators-6qxrx\" (UID: \"e1fc3fae-d6bc-4f40-b01c-06eac982f674\") " pod="openshift-marketplace/certified-operators-6qxrx" Nov 29 05:54:00 crc kubenswrapper[4799]: I1129 05:54:00.404990 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb075ee4-ae05-449f-abac-14ea3f614f2f-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 05:54:00 crc kubenswrapper[4799]: I1129 05:54:00.405703 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1fc3fae-d6bc-4f40-b01c-06eac982f674-utilities\") pod \"certified-operators-6qxrx\" (UID: \"e1fc3fae-d6bc-4f40-b01c-06eac982f674\") " pod="openshift-marketplace/certified-operators-6qxrx" Nov 29 05:54:00 crc kubenswrapper[4799]: I1129 05:54:00.406053 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1fc3fae-d6bc-4f40-b01c-06eac982f674-catalog-content\") pod \"certified-operators-6qxrx\" (UID: \"e1fc3fae-d6bc-4f40-b01c-06eac982f674\") " pod="openshift-marketplace/certified-operators-6qxrx" Nov 29 05:54:00 crc kubenswrapper[4799]: I1129 05:54:00.432487 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6c6n\" (UniqueName: \"kubernetes.io/projected/e1fc3fae-d6bc-4f40-b01c-06eac982f674-kube-api-access-t6c6n\") pod \"certified-operators-6qxrx\" (UID: \"e1fc3fae-d6bc-4f40-b01c-06eac982f674\") " pod="openshift-marketplace/certified-operators-6qxrx" Nov 29 05:54:00 crc kubenswrapper[4799]: I1129 05:54:00.448150 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb075ee4-ae05-449f-abac-14ea3f614f2f-kube-api-access-h972k" (OuterVolumeSpecName: "kube-api-access-h972k") pod "cb075ee4-ae05-449f-abac-14ea3f614f2f" (UID: "cb075ee4-ae05-449f-abac-14ea3f614f2f"). InnerVolumeSpecName "kube-api-access-h972k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:54:00 crc kubenswrapper[4799]: I1129 05:54:00.507335 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h972k\" (UniqueName: \"kubernetes.io/projected/cb075ee4-ae05-449f-abac-14ea3f614f2f-kube-api-access-h972k\") on node \"crc\" DevicePath \"\"" Nov 29 05:54:00 crc kubenswrapper[4799]: I1129 05:54:00.580912 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb075ee4-ae05-449f-abac-14ea3f614f2f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cb075ee4-ae05-449f-abac-14ea3f614f2f" (UID: "cb075ee4-ae05-449f-abac-14ea3f614f2f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:54:00 crc kubenswrapper[4799]: I1129 05:54:00.609048 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb075ee4-ae05-449f-abac-14ea3f614f2f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 05:54:00 crc kubenswrapper[4799]: I1129 05:54:00.664348 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6qxrx" Nov 29 05:54:00 crc kubenswrapper[4799]: I1129 05:54:00.776396 4799 generic.go:334] "Generic (PLEG): container finished" podID="cb075ee4-ae05-449f-abac-14ea3f614f2f" containerID="6641b638b7c7bef08c558fce13fd39a5dfe050050b44603676245206a56ffc00" exitCode=0 Nov 29 05:54:00 crc kubenswrapper[4799]: I1129 05:54:00.776437 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dxhxp" event={"ID":"cb075ee4-ae05-449f-abac-14ea3f614f2f","Type":"ContainerDied","Data":"6641b638b7c7bef08c558fce13fd39a5dfe050050b44603676245206a56ffc00"} Nov 29 05:54:00 crc kubenswrapper[4799]: I1129 05:54:00.776449 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dxhxp" Nov 29 05:54:00 crc kubenswrapper[4799]: I1129 05:54:00.776463 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dxhxp" event={"ID":"cb075ee4-ae05-449f-abac-14ea3f614f2f","Type":"ContainerDied","Data":"1449c52087017efb82822daabee4bf233a96853c4d851986c0a2db1863088174"} Nov 29 05:54:00 crc kubenswrapper[4799]: I1129 05:54:00.776480 4799 scope.go:117] "RemoveContainer" containerID="6641b638b7c7bef08c558fce13fd39a5dfe050050b44603676245206a56ffc00" Nov 29 05:54:00 crc kubenswrapper[4799]: I1129 05:54:00.808465 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dxhxp"] Nov 29 05:54:00 crc kubenswrapper[4799]: I1129 05:54:00.808881 4799 scope.go:117] "RemoveContainer" containerID="397b0bdba6fb759747e2d89ca48e2b4030df290b8216add6e905ad116c1c2ebb" Nov 29 05:54:00 crc kubenswrapper[4799]: I1129 05:54:00.821781 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dxhxp"] Nov 29 05:54:00 crc kubenswrapper[4799]: I1129 05:54:00.867076 4799 scope.go:117] "RemoveContainer" containerID="aff747c64c96600960e92fdf0e7308b360330f95c51345ef48c133fd570d1f27" Nov 29 05:54:00 crc kubenswrapper[4799]: I1129 05:54:00.937650 4799 scope.go:117] "RemoveContainer" containerID="6641b638b7c7bef08c558fce13fd39a5dfe050050b44603676245206a56ffc00" Nov 29 05:54:00 crc kubenswrapper[4799]: E1129 05:54:00.939531 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6641b638b7c7bef08c558fce13fd39a5dfe050050b44603676245206a56ffc00\": container with ID starting with 6641b638b7c7bef08c558fce13fd39a5dfe050050b44603676245206a56ffc00 not found: ID does not exist" containerID="6641b638b7c7bef08c558fce13fd39a5dfe050050b44603676245206a56ffc00" Nov 29 05:54:00 crc kubenswrapper[4799]: I1129 05:54:00.939557 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6641b638b7c7bef08c558fce13fd39a5dfe050050b44603676245206a56ffc00"} err="failed to get container status \"6641b638b7c7bef08c558fce13fd39a5dfe050050b44603676245206a56ffc00\": rpc error: code = NotFound desc = could not find container \"6641b638b7c7bef08c558fce13fd39a5dfe050050b44603676245206a56ffc00\": container with ID starting with 6641b638b7c7bef08c558fce13fd39a5dfe050050b44603676245206a56ffc00 not found: ID does not exist" Nov 29 05:54:00 crc kubenswrapper[4799]: I1129 05:54:00.939578 4799 scope.go:117] "RemoveContainer" containerID="397b0bdba6fb759747e2d89ca48e2b4030df290b8216add6e905ad116c1c2ebb" Nov 29 05:54:00 crc kubenswrapper[4799]: E1129 05:54:00.940033 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"397b0bdba6fb759747e2d89ca48e2b4030df290b8216add6e905ad116c1c2ebb\": container with ID starting with 397b0bdba6fb759747e2d89ca48e2b4030df290b8216add6e905ad116c1c2ebb not found: ID does not exist" containerID="397b0bdba6fb759747e2d89ca48e2b4030df290b8216add6e905ad116c1c2ebb" Nov 29 05:54:00 crc kubenswrapper[4799]: I1129 05:54:00.940052 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"397b0bdba6fb759747e2d89ca48e2b4030df290b8216add6e905ad116c1c2ebb"} err="failed to get container status \"397b0bdba6fb759747e2d89ca48e2b4030df290b8216add6e905ad116c1c2ebb\": rpc error: code = NotFound desc = could not find container \"397b0bdba6fb759747e2d89ca48e2b4030df290b8216add6e905ad116c1c2ebb\": container with ID starting with 397b0bdba6fb759747e2d89ca48e2b4030df290b8216add6e905ad116c1c2ebb not found: ID does not exist" Nov 29 05:54:00 crc kubenswrapper[4799]: I1129 05:54:00.940078 4799 scope.go:117] "RemoveContainer" containerID="aff747c64c96600960e92fdf0e7308b360330f95c51345ef48c133fd570d1f27" Nov 29 05:54:00 crc kubenswrapper[4799]: E1129 05:54:00.941184 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aff747c64c96600960e92fdf0e7308b360330f95c51345ef48c133fd570d1f27\": container with ID starting with aff747c64c96600960e92fdf0e7308b360330f95c51345ef48c133fd570d1f27 not found: ID does not exist" containerID="aff747c64c96600960e92fdf0e7308b360330f95c51345ef48c133fd570d1f27" Nov 29 05:54:00 crc kubenswrapper[4799]: I1129 05:54:00.941201 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aff747c64c96600960e92fdf0e7308b360330f95c51345ef48c133fd570d1f27"} err="failed to get container status \"aff747c64c96600960e92fdf0e7308b360330f95c51345ef48c133fd570d1f27\": rpc error: code = NotFound desc = could not find container \"aff747c64c96600960e92fdf0e7308b360330f95c51345ef48c133fd570d1f27\": container with ID starting with aff747c64c96600960e92fdf0e7308b360330f95c51345ef48c133fd570d1f27 not found: ID does not exist" Nov 29 05:54:01 crc kubenswrapper[4799]: I1129 05:54:01.209080 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6qxrx"] Nov 29 05:54:01 crc kubenswrapper[4799]: W1129 05:54:01.213055 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode1fc3fae_d6bc_4f40_b01c_06eac982f674.slice/crio-df7e40a3322453e433b1c7578acc10a8bcba78cfc2471b57474d356c3df3e73b WatchSource:0}: Error finding container df7e40a3322453e433b1c7578acc10a8bcba78cfc2471b57474d356c3df3e73b: Status 404 returned error can't find the container with id df7e40a3322453e433b1c7578acc10a8bcba78cfc2471b57474d356c3df3e73b Nov 29 05:54:01 crc kubenswrapper[4799]: I1129 05:54:01.791585 4799 generic.go:334] "Generic (PLEG): container finished" podID="e1fc3fae-d6bc-4f40-b01c-06eac982f674" containerID="9ee7b17bd57d8dbec19cbf47fa9a95f885d7241a6696993bdb74663ed6a53de3" exitCode=0 Nov 29 05:54:01 crc kubenswrapper[4799]: I1129 05:54:01.791672 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6qxrx" event={"ID":"e1fc3fae-d6bc-4f40-b01c-06eac982f674","Type":"ContainerDied","Data":"9ee7b17bd57d8dbec19cbf47fa9a95f885d7241a6696993bdb74663ed6a53de3"} Nov 29 05:54:01 crc kubenswrapper[4799]: I1129 05:54:01.791998 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6qxrx" event={"ID":"e1fc3fae-d6bc-4f40-b01c-06eac982f674","Type":"ContainerStarted","Data":"df7e40a3322453e433b1c7578acc10a8bcba78cfc2471b57474d356c3df3e73b"} Nov 29 05:54:02 crc kubenswrapper[4799]: I1129 05:54:02.670644 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb075ee4-ae05-449f-abac-14ea3f614f2f" path="/var/lib/kubelet/pods/cb075ee4-ae05-449f-abac-14ea3f614f2f/volumes" Nov 29 05:54:02 crc kubenswrapper[4799]: I1129 05:54:02.804911 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6qxrx" event={"ID":"e1fc3fae-d6bc-4f40-b01c-06eac982f674","Type":"ContainerStarted","Data":"3d82a3e783f165ca00108bbdab5d571062b32f310d0f0c71625b9369e64b7efc"} Nov 29 05:54:03 crc kubenswrapper[4799]: I1129 05:54:03.814467 4799 generic.go:334] "Generic (PLEG): container finished" podID="e1fc3fae-d6bc-4f40-b01c-06eac982f674" containerID="3d82a3e783f165ca00108bbdab5d571062b32f310d0f0c71625b9369e64b7efc" exitCode=0 Nov 29 05:54:03 crc kubenswrapper[4799]: I1129 05:54:03.814520 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6qxrx" event={"ID":"e1fc3fae-d6bc-4f40-b01c-06eac982f674","Type":"ContainerDied","Data":"3d82a3e783f165ca00108bbdab5d571062b32f310d0f0c71625b9369e64b7efc"} Nov 29 05:54:04 crc kubenswrapper[4799]: I1129 05:54:04.827706 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6qxrx" event={"ID":"e1fc3fae-d6bc-4f40-b01c-06eac982f674","Type":"ContainerStarted","Data":"5fcd482ed358db5494daea12d67bcc308eafa40164df97babdd8bf2edb247137"} Nov 29 05:54:04 crc kubenswrapper[4799]: I1129 05:54:04.844602 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6qxrx" podStartSLOduration=2.434866426 podStartE2EDuration="4.844580107s" podCreationTimestamp="2025-11-29 05:54:00 +0000 UTC" firstStartedPulling="2025-11-29 05:54:01.802771339 +0000 UTC m=+4517.445701739" lastFinishedPulling="2025-11-29 05:54:04.21248502 +0000 UTC m=+4519.855415420" observedRunningTime="2025-11-29 05:54:04.844079245 +0000 UTC m=+4520.487009645" watchObservedRunningTime="2025-11-29 05:54:04.844580107 +0000 UTC m=+4520.487510507" Nov 29 05:54:07 crc kubenswrapper[4799]: I1129 05:54:07.659612 4799 scope.go:117] "RemoveContainer" containerID="b560e06b6de1edd9a7f603575cb604d0893a8a24eda3e04d2446b813f4e889f8" Nov 29 05:54:07 crc kubenswrapper[4799]: E1129 05:54:07.660478 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:54:10 crc kubenswrapper[4799]: I1129 05:54:10.668454 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6qxrx" Nov 29 05:54:10 crc kubenswrapper[4799]: I1129 05:54:10.669030 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6qxrx" Nov 29 05:54:10 crc kubenswrapper[4799]: I1129 05:54:10.739960 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6qxrx" Nov 29 05:54:10 crc kubenswrapper[4799]: I1129 05:54:10.920028 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6qxrx" Nov 29 05:54:10 crc kubenswrapper[4799]: I1129 05:54:10.976414 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6qxrx"] Nov 29 05:54:12 crc kubenswrapper[4799]: I1129 05:54:12.903261 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6qxrx" podUID="e1fc3fae-d6bc-4f40-b01c-06eac982f674" containerName="registry-server" containerID="cri-o://5fcd482ed358db5494daea12d67bcc308eafa40164df97babdd8bf2edb247137" gracePeriod=2 Nov 29 05:54:13 crc kubenswrapper[4799]: I1129 05:54:13.443647 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6qxrx" Nov 29 05:54:13 crc kubenswrapper[4799]: I1129 05:54:13.569588 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t6c6n\" (UniqueName: \"kubernetes.io/projected/e1fc3fae-d6bc-4f40-b01c-06eac982f674-kube-api-access-t6c6n\") pod \"e1fc3fae-d6bc-4f40-b01c-06eac982f674\" (UID: \"e1fc3fae-d6bc-4f40-b01c-06eac982f674\") " Nov 29 05:54:13 crc kubenswrapper[4799]: I1129 05:54:13.569701 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1fc3fae-d6bc-4f40-b01c-06eac982f674-catalog-content\") pod \"e1fc3fae-d6bc-4f40-b01c-06eac982f674\" (UID: \"e1fc3fae-d6bc-4f40-b01c-06eac982f674\") " Nov 29 05:54:13 crc kubenswrapper[4799]: I1129 05:54:13.569733 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1fc3fae-d6bc-4f40-b01c-06eac982f674-utilities\") pod \"e1fc3fae-d6bc-4f40-b01c-06eac982f674\" (UID: \"e1fc3fae-d6bc-4f40-b01c-06eac982f674\") " Nov 29 05:54:13 crc kubenswrapper[4799]: I1129 05:54:13.571057 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1fc3fae-d6bc-4f40-b01c-06eac982f674-utilities" (OuterVolumeSpecName: "utilities") pod "e1fc3fae-d6bc-4f40-b01c-06eac982f674" (UID: "e1fc3fae-d6bc-4f40-b01c-06eac982f674"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:54:13 crc kubenswrapper[4799]: I1129 05:54:13.584025 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1fc3fae-d6bc-4f40-b01c-06eac982f674-kube-api-access-t6c6n" (OuterVolumeSpecName: "kube-api-access-t6c6n") pod "e1fc3fae-d6bc-4f40-b01c-06eac982f674" (UID: "e1fc3fae-d6bc-4f40-b01c-06eac982f674"). InnerVolumeSpecName "kube-api-access-t6c6n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:54:13 crc kubenswrapper[4799]: I1129 05:54:13.627000 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1fc3fae-d6bc-4f40-b01c-06eac982f674-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e1fc3fae-d6bc-4f40-b01c-06eac982f674" (UID: "e1fc3fae-d6bc-4f40-b01c-06eac982f674"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:54:13 crc kubenswrapper[4799]: I1129 05:54:13.671846 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t6c6n\" (UniqueName: \"kubernetes.io/projected/e1fc3fae-d6bc-4f40-b01c-06eac982f674-kube-api-access-t6c6n\") on node \"crc\" DevicePath \"\"" Nov 29 05:54:13 crc kubenswrapper[4799]: I1129 05:54:13.672041 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1fc3fae-d6bc-4f40-b01c-06eac982f674-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 05:54:13 crc kubenswrapper[4799]: I1129 05:54:13.672253 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1fc3fae-d6bc-4f40-b01c-06eac982f674-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 05:54:13 crc kubenswrapper[4799]: I1129 05:54:13.913608 4799 generic.go:334] "Generic (PLEG): container finished" podID="e1fc3fae-d6bc-4f40-b01c-06eac982f674" containerID="5fcd482ed358db5494daea12d67bcc308eafa40164df97babdd8bf2edb247137" exitCode=0 Nov 29 05:54:13 crc kubenswrapper[4799]: I1129 05:54:13.913696 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6qxrx" Nov 29 05:54:13 crc kubenswrapper[4799]: I1129 05:54:13.913687 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6qxrx" event={"ID":"e1fc3fae-d6bc-4f40-b01c-06eac982f674","Type":"ContainerDied","Data":"5fcd482ed358db5494daea12d67bcc308eafa40164df97babdd8bf2edb247137"} Nov 29 05:54:13 crc kubenswrapper[4799]: I1129 05:54:13.915175 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6qxrx" event={"ID":"e1fc3fae-d6bc-4f40-b01c-06eac982f674","Type":"ContainerDied","Data":"df7e40a3322453e433b1c7578acc10a8bcba78cfc2471b57474d356c3df3e73b"} Nov 29 05:54:13 crc kubenswrapper[4799]: I1129 05:54:13.915202 4799 scope.go:117] "RemoveContainer" containerID="5fcd482ed358db5494daea12d67bcc308eafa40164df97babdd8bf2edb247137" Nov 29 05:54:13 crc kubenswrapper[4799]: I1129 05:54:13.940342 4799 scope.go:117] "RemoveContainer" containerID="3d82a3e783f165ca00108bbdab5d571062b32f310d0f0c71625b9369e64b7efc" Nov 29 05:54:13 crc kubenswrapper[4799]: I1129 05:54:13.954069 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6qxrx"] Nov 29 05:54:13 crc kubenswrapper[4799]: I1129 05:54:13.961742 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6qxrx"] Nov 29 05:54:13 crc kubenswrapper[4799]: I1129 05:54:13.970428 4799 scope.go:117] "RemoveContainer" containerID="9ee7b17bd57d8dbec19cbf47fa9a95f885d7241a6696993bdb74663ed6a53de3" Nov 29 05:54:14 crc kubenswrapper[4799]: I1129 05:54:14.015217 4799 scope.go:117] "RemoveContainer" containerID="5fcd482ed358db5494daea12d67bcc308eafa40164df97babdd8bf2edb247137" Nov 29 05:54:14 crc kubenswrapper[4799]: E1129 05:54:14.015700 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5fcd482ed358db5494daea12d67bcc308eafa40164df97babdd8bf2edb247137\": container with ID starting with 5fcd482ed358db5494daea12d67bcc308eafa40164df97babdd8bf2edb247137 not found: ID does not exist" containerID="5fcd482ed358db5494daea12d67bcc308eafa40164df97babdd8bf2edb247137" Nov 29 05:54:14 crc kubenswrapper[4799]: I1129 05:54:14.015734 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5fcd482ed358db5494daea12d67bcc308eafa40164df97babdd8bf2edb247137"} err="failed to get container status \"5fcd482ed358db5494daea12d67bcc308eafa40164df97babdd8bf2edb247137\": rpc error: code = NotFound desc = could not find container \"5fcd482ed358db5494daea12d67bcc308eafa40164df97babdd8bf2edb247137\": container with ID starting with 5fcd482ed358db5494daea12d67bcc308eafa40164df97babdd8bf2edb247137 not found: ID does not exist" Nov 29 05:54:14 crc kubenswrapper[4799]: I1129 05:54:14.015752 4799 scope.go:117] "RemoveContainer" containerID="3d82a3e783f165ca00108bbdab5d571062b32f310d0f0c71625b9369e64b7efc" Nov 29 05:54:14 crc kubenswrapper[4799]: E1129 05:54:14.017535 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d82a3e783f165ca00108bbdab5d571062b32f310d0f0c71625b9369e64b7efc\": container with ID starting with 3d82a3e783f165ca00108bbdab5d571062b32f310d0f0c71625b9369e64b7efc not found: ID does not exist" containerID="3d82a3e783f165ca00108bbdab5d571062b32f310d0f0c71625b9369e64b7efc" Nov 29 05:54:14 crc kubenswrapper[4799]: I1129 05:54:14.017588 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d82a3e783f165ca00108bbdab5d571062b32f310d0f0c71625b9369e64b7efc"} err="failed to get container status \"3d82a3e783f165ca00108bbdab5d571062b32f310d0f0c71625b9369e64b7efc\": rpc error: code = NotFound desc = could not find container \"3d82a3e783f165ca00108bbdab5d571062b32f310d0f0c71625b9369e64b7efc\": container with ID starting with 3d82a3e783f165ca00108bbdab5d571062b32f310d0f0c71625b9369e64b7efc not found: ID does not exist" Nov 29 05:54:14 crc kubenswrapper[4799]: I1129 05:54:14.017621 4799 scope.go:117] "RemoveContainer" containerID="9ee7b17bd57d8dbec19cbf47fa9a95f885d7241a6696993bdb74663ed6a53de3" Nov 29 05:54:14 crc kubenswrapper[4799]: E1129 05:54:14.018069 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ee7b17bd57d8dbec19cbf47fa9a95f885d7241a6696993bdb74663ed6a53de3\": container with ID starting with 9ee7b17bd57d8dbec19cbf47fa9a95f885d7241a6696993bdb74663ed6a53de3 not found: ID does not exist" containerID="9ee7b17bd57d8dbec19cbf47fa9a95f885d7241a6696993bdb74663ed6a53de3" Nov 29 05:54:14 crc kubenswrapper[4799]: I1129 05:54:14.018103 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ee7b17bd57d8dbec19cbf47fa9a95f885d7241a6696993bdb74663ed6a53de3"} err="failed to get container status \"9ee7b17bd57d8dbec19cbf47fa9a95f885d7241a6696993bdb74663ed6a53de3\": rpc error: code = NotFound desc = could not find container \"9ee7b17bd57d8dbec19cbf47fa9a95f885d7241a6696993bdb74663ed6a53de3\": container with ID starting with 9ee7b17bd57d8dbec19cbf47fa9a95f885d7241a6696993bdb74663ed6a53de3 not found: ID does not exist" Nov 29 05:54:14 crc kubenswrapper[4799]: I1129 05:54:14.677990 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1fc3fae-d6bc-4f40-b01c-06eac982f674" path="/var/lib/kubelet/pods/e1fc3fae-d6bc-4f40-b01c-06eac982f674/volumes" Nov 29 05:54:22 crc kubenswrapper[4799]: I1129 05:54:22.659077 4799 scope.go:117] "RemoveContainer" containerID="b560e06b6de1edd9a7f603575cb604d0893a8a24eda3e04d2446b813f4e889f8" Nov 29 05:54:22 crc kubenswrapper[4799]: E1129 05:54:22.660011 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:54:33 crc kubenswrapper[4799]: I1129 05:54:33.658750 4799 scope.go:117] "RemoveContainer" containerID="b560e06b6de1edd9a7f603575cb604d0893a8a24eda3e04d2446b813f4e889f8" Nov 29 05:54:33 crc kubenswrapper[4799]: E1129 05:54:33.659503 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:54:38 crc kubenswrapper[4799]: I1129 05:54:38.125800 4799 generic.go:334] "Generic (PLEG): container finished" podID="474c86be-9e78-48eb-b5b0-60db74e967e3" containerID="499d0333bd30be0734900a58a7daf348badb369ce4c1cb2598b375bd380a1b12" exitCode=0 Nov 29 05:54:38 crc kubenswrapper[4799]: I1129 05:54:38.125902 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t9ms2/crc-debug-dndqf" event={"ID":"474c86be-9e78-48eb-b5b0-60db74e967e3","Type":"ContainerDied","Data":"499d0333bd30be0734900a58a7daf348badb369ce4c1cb2598b375bd380a1b12"} Nov 29 05:54:39 crc kubenswrapper[4799]: I1129 05:54:39.250481 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t9ms2/crc-debug-dndqf" Nov 29 05:54:39 crc kubenswrapper[4799]: I1129 05:54:39.285364 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-t9ms2/crc-debug-dndqf"] Nov 29 05:54:39 crc kubenswrapper[4799]: I1129 05:54:39.293866 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-t9ms2/crc-debug-dndqf"] Nov 29 05:54:39 crc kubenswrapper[4799]: I1129 05:54:39.428524 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrpx5\" (UniqueName: \"kubernetes.io/projected/474c86be-9e78-48eb-b5b0-60db74e967e3-kube-api-access-jrpx5\") pod \"474c86be-9e78-48eb-b5b0-60db74e967e3\" (UID: \"474c86be-9e78-48eb-b5b0-60db74e967e3\") " Nov 29 05:54:39 crc kubenswrapper[4799]: I1129 05:54:39.428587 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/474c86be-9e78-48eb-b5b0-60db74e967e3-host\") pod \"474c86be-9e78-48eb-b5b0-60db74e967e3\" (UID: \"474c86be-9e78-48eb-b5b0-60db74e967e3\") " Nov 29 05:54:39 crc kubenswrapper[4799]: I1129 05:54:39.428697 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/474c86be-9e78-48eb-b5b0-60db74e967e3-host" (OuterVolumeSpecName: "host") pod "474c86be-9e78-48eb-b5b0-60db74e967e3" (UID: "474c86be-9e78-48eb-b5b0-60db74e967e3"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 05:54:39 crc kubenswrapper[4799]: I1129 05:54:39.429177 4799 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/474c86be-9e78-48eb-b5b0-60db74e967e3-host\") on node \"crc\" DevicePath \"\"" Nov 29 05:54:39 crc kubenswrapper[4799]: I1129 05:54:39.433449 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/474c86be-9e78-48eb-b5b0-60db74e967e3-kube-api-access-jrpx5" (OuterVolumeSpecName: "kube-api-access-jrpx5") pod "474c86be-9e78-48eb-b5b0-60db74e967e3" (UID: "474c86be-9e78-48eb-b5b0-60db74e967e3"). InnerVolumeSpecName "kube-api-access-jrpx5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:54:39 crc kubenswrapper[4799]: I1129 05:54:39.530260 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrpx5\" (UniqueName: \"kubernetes.io/projected/474c86be-9e78-48eb-b5b0-60db74e967e3-kube-api-access-jrpx5\") on node \"crc\" DevicePath \"\"" Nov 29 05:54:40 crc kubenswrapper[4799]: I1129 05:54:40.146083 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ef0b744ea4f6579b61788bfaef631d414b3f2d29987941ed66f0fbe9df9ad447" Nov 29 05:54:40 crc kubenswrapper[4799]: I1129 05:54:40.146115 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t9ms2/crc-debug-dndqf" Nov 29 05:54:40 crc kubenswrapper[4799]: I1129 05:54:40.484293 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-t9ms2/crc-debug-gjrbs"] Nov 29 05:54:40 crc kubenswrapper[4799]: E1129 05:54:40.485591 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb075ee4-ae05-449f-abac-14ea3f614f2f" containerName="extract-utilities" Nov 29 05:54:40 crc kubenswrapper[4799]: I1129 05:54:40.485699 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb075ee4-ae05-449f-abac-14ea3f614f2f" containerName="extract-utilities" Nov 29 05:54:40 crc kubenswrapper[4799]: E1129 05:54:40.485824 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="474c86be-9e78-48eb-b5b0-60db74e967e3" containerName="container-00" Nov 29 05:54:40 crc kubenswrapper[4799]: I1129 05:54:40.485908 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="474c86be-9e78-48eb-b5b0-60db74e967e3" containerName="container-00" Nov 29 05:54:40 crc kubenswrapper[4799]: E1129 05:54:40.486003 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb075ee4-ae05-449f-abac-14ea3f614f2f" containerName="registry-server" Nov 29 05:54:40 crc kubenswrapper[4799]: I1129 05:54:40.486080 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb075ee4-ae05-449f-abac-14ea3f614f2f" containerName="registry-server" Nov 29 05:54:40 crc kubenswrapper[4799]: E1129 05:54:40.486163 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1fc3fae-d6bc-4f40-b01c-06eac982f674" containerName="extract-utilities" Nov 29 05:54:40 crc kubenswrapper[4799]: I1129 05:54:40.486240 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1fc3fae-d6bc-4f40-b01c-06eac982f674" containerName="extract-utilities" Nov 29 05:54:40 crc kubenswrapper[4799]: E1129 05:54:40.486322 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1fc3fae-d6bc-4f40-b01c-06eac982f674" containerName="extract-content" Nov 29 05:54:40 crc kubenswrapper[4799]: I1129 05:54:40.486393 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1fc3fae-d6bc-4f40-b01c-06eac982f674" containerName="extract-content" Nov 29 05:54:40 crc kubenswrapper[4799]: E1129 05:54:40.486479 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1fc3fae-d6bc-4f40-b01c-06eac982f674" containerName="registry-server" Nov 29 05:54:40 crc kubenswrapper[4799]: I1129 05:54:40.486556 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1fc3fae-d6bc-4f40-b01c-06eac982f674" containerName="registry-server" Nov 29 05:54:40 crc kubenswrapper[4799]: E1129 05:54:40.486631 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb075ee4-ae05-449f-abac-14ea3f614f2f" containerName="extract-content" Nov 29 05:54:40 crc kubenswrapper[4799]: I1129 05:54:40.486704 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb075ee4-ae05-449f-abac-14ea3f614f2f" containerName="extract-content" Nov 29 05:54:40 crc kubenswrapper[4799]: I1129 05:54:40.487069 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb075ee4-ae05-449f-abac-14ea3f614f2f" containerName="registry-server" Nov 29 05:54:40 crc kubenswrapper[4799]: I1129 05:54:40.487198 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1fc3fae-d6bc-4f40-b01c-06eac982f674" containerName="registry-server" Nov 29 05:54:40 crc kubenswrapper[4799]: I1129 05:54:40.487298 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="474c86be-9e78-48eb-b5b0-60db74e967e3" containerName="container-00" Nov 29 05:54:40 crc kubenswrapper[4799]: I1129 05:54:40.488224 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t9ms2/crc-debug-gjrbs" Nov 29 05:54:40 crc kubenswrapper[4799]: I1129 05:54:40.650748 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ca56befe-f116-4bdc-b657-d67337e4a774-host\") pod \"crc-debug-gjrbs\" (UID: \"ca56befe-f116-4bdc-b657-d67337e4a774\") " pod="openshift-must-gather-t9ms2/crc-debug-gjrbs" Nov 29 05:54:40 crc kubenswrapper[4799]: I1129 05:54:40.650935 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcq8g\" (UniqueName: \"kubernetes.io/projected/ca56befe-f116-4bdc-b657-d67337e4a774-kube-api-access-gcq8g\") pod \"crc-debug-gjrbs\" (UID: \"ca56befe-f116-4bdc-b657-d67337e4a774\") " pod="openshift-must-gather-t9ms2/crc-debug-gjrbs" Nov 29 05:54:40 crc kubenswrapper[4799]: I1129 05:54:40.669474 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="474c86be-9e78-48eb-b5b0-60db74e967e3" path="/var/lib/kubelet/pods/474c86be-9e78-48eb-b5b0-60db74e967e3/volumes" Nov 29 05:54:40 crc kubenswrapper[4799]: I1129 05:54:40.752907 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ca56befe-f116-4bdc-b657-d67337e4a774-host\") pod \"crc-debug-gjrbs\" (UID: \"ca56befe-f116-4bdc-b657-d67337e4a774\") " pod="openshift-must-gather-t9ms2/crc-debug-gjrbs" Nov 29 05:54:40 crc kubenswrapper[4799]: I1129 05:54:40.753276 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ca56befe-f116-4bdc-b657-d67337e4a774-host\") pod \"crc-debug-gjrbs\" (UID: \"ca56befe-f116-4bdc-b657-d67337e4a774\") " pod="openshift-must-gather-t9ms2/crc-debug-gjrbs" Nov 29 05:54:40 crc kubenswrapper[4799]: I1129 05:54:40.753554 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcq8g\" (UniqueName: \"kubernetes.io/projected/ca56befe-f116-4bdc-b657-d67337e4a774-kube-api-access-gcq8g\") pod \"crc-debug-gjrbs\" (UID: \"ca56befe-f116-4bdc-b657-d67337e4a774\") " pod="openshift-must-gather-t9ms2/crc-debug-gjrbs" Nov 29 05:54:40 crc kubenswrapper[4799]: I1129 05:54:40.774666 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcq8g\" (UniqueName: \"kubernetes.io/projected/ca56befe-f116-4bdc-b657-d67337e4a774-kube-api-access-gcq8g\") pod \"crc-debug-gjrbs\" (UID: \"ca56befe-f116-4bdc-b657-d67337e4a774\") " pod="openshift-must-gather-t9ms2/crc-debug-gjrbs" Nov 29 05:54:40 crc kubenswrapper[4799]: I1129 05:54:40.808517 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t9ms2/crc-debug-gjrbs" Nov 29 05:54:41 crc kubenswrapper[4799]: I1129 05:54:41.162086 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t9ms2/crc-debug-gjrbs" event={"ID":"ca56befe-f116-4bdc-b657-d67337e4a774","Type":"ContainerStarted","Data":"cd5f1d2858843448767b327c15763b00e79599889269e9b82425990c734cba43"} Nov 29 05:54:41 crc kubenswrapper[4799]: I1129 05:54:41.162557 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t9ms2/crc-debug-gjrbs" event={"ID":"ca56befe-f116-4bdc-b657-d67337e4a774","Type":"ContainerStarted","Data":"0f13d71e64d234da954915019b05de2b7327caf2b49dbefd030e143f8146f34c"} Nov 29 05:54:41 crc kubenswrapper[4799]: I1129 05:54:41.178892 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-t9ms2/crc-debug-gjrbs" podStartSLOduration=1.17886861 podStartE2EDuration="1.17886861s" podCreationTimestamp="2025-11-29 05:54:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 05:54:41.176023402 +0000 UTC m=+4556.818953802" watchObservedRunningTime="2025-11-29 05:54:41.17886861 +0000 UTC m=+4556.821799010" Nov 29 05:54:42 crc kubenswrapper[4799]: I1129 05:54:42.181710 4799 generic.go:334] "Generic (PLEG): container finished" podID="ca56befe-f116-4bdc-b657-d67337e4a774" containerID="cd5f1d2858843448767b327c15763b00e79599889269e9b82425990c734cba43" exitCode=0 Nov 29 05:54:42 crc kubenswrapper[4799]: I1129 05:54:42.181752 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t9ms2/crc-debug-gjrbs" event={"ID":"ca56befe-f116-4bdc-b657-d67337e4a774","Type":"ContainerDied","Data":"cd5f1d2858843448767b327c15763b00e79599889269e9b82425990c734cba43"} Nov 29 05:54:43 crc kubenswrapper[4799]: I1129 05:54:43.286464 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t9ms2/crc-debug-gjrbs" Nov 29 05:54:43 crc kubenswrapper[4799]: I1129 05:54:43.317665 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-t9ms2/crc-debug-gjrbs"] Nov 29 05:54:43 crc kubenswrapper[4799]: I1129 05:54:43.325530 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-t9ms2/crc-debug-gjrbs"] Nov 29 05:54:43 crc kubenswrapper[4799]: I1129 05:54:43.403707 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ca56befe-f116-4bdc-b657-d67337e4a774-host\") pod \"ca56befe-f116-4bdc-b657-d67337e4a774\" (UID: \"ca56befe-f116-4bdc-b657-d67337e4a774\") " Nov 29 05:54:43 crc kubenswrapper[4799]: I1129 05:54:43.403785 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ca56befe-f116-4bdc-b657-d67337e4a774-host" (OuterVolumeSpecName: "host") pod "ca56befe-f116-4bdc-b657-d67337e4a774" (UID: "ca56befe-f116-4bdc-b657-d67337e4a774"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 05:54:43 crc kubenswrapper[4799]: I1129 05:54:43.404263 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gcq8g\" (UniqueName: \"kubernetes.io/projected/ca56befe-f116-4bdc-b657-d67337e4a774-kube-api-access-gcq8g\") pod \"ca56befe-f116-4bdc-b657-d67337e4a774\" (UID: \"ca56befe-f116-4bdc-b657-d67337e4a774\") " Nov 29 05:54:43 crc kubenswrapper[4799]: I1129 05:54:43.404840 4799 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ca56befe-f116-4bdc-b657-d67337e4a774-host\") on node \"crc\" DevicePath \"\"" Nov 29 05:54:43 crc kubenswrapper[4799]: I1129 05:54:43.410746 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca56befe-f116-4bdc-b657-d67337e4a774-kube-api-access-gcq8g" (OuterVolumeSpecName: "kube-api-access-gcq8g") pod "ca56befe-f116-4bdc-b657-d67337e4a774" (UID: "ca56befe-f116-4bdc-b657-d67337e4a774"). InnerVolumeSpecName "kube-api-access-gcq8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:54:43 crc kubenswrapper[4799]: I1129 05:54:43.507222 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gcq8g\" (UniqueName: \"kubernetes.io/projected/ca56befe-f116-4bdc-b657-d67337e4a774-kube-api-access-gcq8g\") on node \"crc\" DevicePath \"\"" Nov 29 05:54:44 crc kubenswrapper[4799]: I1129 05:54:44.200112 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0f13d71e64d234da954915019b05de2b7327caf2b49dbefd030e143f8146f34c" Nov 29 05:54:44 crc kubenswrapper[4799]: I1129 05:54:44.200369 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t9ms2/crc-debug-gjrbs" Nov 29 05:54:44 crc kubenswrapper[4799]: I1129 05:54:44.547251 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-t9ms2/crc-debug-t7hcg"] Nov 29 05:54:44 crc kubenswrapper[4799]: E1129 05:54:44.548376 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca56befe-f116-4bdc-b657-d67337e4a774" containerName="container-00" Nov 29 05:54:44 crc kubenswrapper[4799]: I1129 05:54:44.548470 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca56befe-f116-4bdc-b657-d67337e4a774" containerName="container-00" Nov 29 05:54:44 crc kubenswrapper[4799]: I1129 05:54:44.548768 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca56befe-f116-4bdc-b657-d67337e4a774" containerName="container-00" Nov 29 05:54:44 crc kubenswrapper[4799]: I1129 05:54:44.549701 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t9ms2/crc-debug-t7hcg" Nov 29 05:54:44 crc kubenswrapper[4799]: I1129 05:54:44.664586 4799 scope.go:117] "RemoveContainer" containerID="b560e06b6de1edd9a7f603575cb604d0893a8a24eda3e04d2446b813f4e889f8" Nov 29 05:54:44 crc kubenswrapper[4799]: E1129 05:54:44.664935 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:54:44 crc kubenswrapper[4799]: I1129 05:54:44.670682 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca56befe-f116-4bdc-b657-d67337e4a774" path="/var/lib/kubelet/pods/ca56befe-f116-4bdc-b657-d67337e4a774/volumes" Nov 29 05:54:44 crc kubenswrapper[4799]: I1129 05:54:44.729753 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/16e3527e-df72-474b-9905-e770818caa33-host\") pod \"crc-debug-t7hcg\" (UID: \"16e3527e-df72-474b-9905-e770818caa33\") " pod="openshift-must-gather-t9ms2/crc-debug-t7hcg" Nov 29 05:54:44 crc kubenswrapper[4799]: I1129 05:54:44.729822 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rd744\" (UniqueName: \"kubernetes.io/projected/16e3527e-df72-474b-9905-e770818caa33-kube-api-access-rd744\") pod \"crc-debug-t7hcg\" (UID: \"16e3527e-df72-474b-9905-e770818caa33\") " pod="openshift-must-gather-t9ms2/crc-debug-t7hcg" Nov 29 05:54:44 crc kubenswrapper[4799]: I1129 05:54:44.831125 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/16e3527e-df72-474b-9905-e770818caa33-host\") pod \"crc-debug-t7hcg\" (UID: \"16e3527e-df72-474b-9905-e770818caa33\") " pod="openshift-must-gather-t9ms2/crc-debug-t7hcg" Nov 29 05:54:44 crc kubenswrapper[4799]: I1129 05:54:44.831181 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rd744\" (UniqueName: \"kubernetes.io/projected/16e3527e-df72-474b-9905-e770818caa33-kube-api-access-rd744\") pod \"crc-debug-t7hcg\" (UID: \"16e3527e-df72-474b-9905-e770818caa33\") " pod="openshift-must-gather-t9ms2/crc-debug-t7hcg" Nov 29 05:54:44 crc kubenswrapper[4799]: I1129 05:54:44.831855 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/16e3527e-df72-474b-9905-e770818caa33-host\") pod \"crc-debug-t7hcg\" (UID: \"16e3527e-df72-474b-9905-e770818caa33\") " pod="openshift-must-gather-t9ms2/crc-debug-t7hcg" Nov 29 05:54:44 crc kubenswrapper[4799]: I1129 05:54:44.853423 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rd744\" (UniqueName: \"kubernetes.io/projected/16e3527e-df72-474b-9905-e770818caa33-kube-api-access-rd744\") pod \"crc-debug-t7hcg\" (UID: \"16e3527e-df72-474b-9905-e770818caa33\") " pod="openshift-must-gather-t9ms2/crc-debug-t7hcg" Nov 29 05:54:44 crc kubenswrapper[4799]: I1129 05:54:44.869447 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t9ms2/crc-debug-t7hcg" Nov 29 05:54:44 crc kubenswrapper[4799]: W1129 05:54:44.915216 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod16e3527e_df72_474b_9905_e770818caa33.slice/crio-077f3eee97dbe512921d5b5965271f94abd529560acfcfb674695a9e87ab596d WatchSource:0}: Error finding container 077f3eee97dbe512921d5b5965271f94abd529560acfcfb674695a9e87ab596d: Status 404 returned error can't find the container with id 077f3eee97dbe512921d5b5965271f94abd529560acfcfb674695a9e87ab596d Nov 29 05:54:45 crc kubenswrapper[4799]: I1129 05:54:45.208273 4799 generic.go:334] "Generic (PLEG): container finished" podID="16e3527e-df72-474b-9905-e770818caa33" containerID="ec98bac390f9126125f67ab591d022819d6a077ec906b1149d8ccbb45f9c7dd8" exitCode=0 Nov 29 05:54:45 crc kubenswrapper[4799]: I1129 05:54:45.208338 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t9ms2/crc-debug-t7hcg" event={"ID":"16e3527e-df72-474b-9905-e770818caa33","Type":"ContainerDied","Data":"ec98bac390f9126125f67ab591d022819d6a077ec906b1149d8ccbb45f9c7dd8"} Nov 29 05:54:45 crc kubenswrapper[4799]: I1129 05:54:45.208606 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t9ms2/crc-debug-t7hcg" event={"ID":"16e3527e-df72-474b-9905-e770818caa33","Type":"ContainerStarted","Data":"077f3eee97dbe512921d5b5965271f94abd529560acfcfb674695a9e87ab596d"} Nov 29 05:54:45 crc kubenswrapper[4799]: I1129 05:54:45.242091 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-t9ms2/crc-debug-t7hcg"] Nov 29 05:54:45 crc kubenswrapper[4799]: I1129 05:54:45.250351 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-t9ms2/crc-debug-t7hcg"] Nov 29 05:54:46 crc kubenswrapper[4799]: I1129 05:54:46.309107 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t9ms2/crc-debug-t7hcg" Nov 29 05:54:46 crc kubenswrapper[4799]: I1129 05:54:46.458996 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/16e3527e-df72-474b-9905-e770818caa33-host\") pod \"16e3527e-df72-474b-9905-e770818caa33\" (UID: \"16e3527e-df72-474b-9905-e770818caa33\") " Nov 29 05:54:46 crc kubenswrapper[4799]: I1129 05:54:46.459223 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/16e3527e-df72-474b-9905-e770818caa33-host" (OuterVolumeSpecName: "host") pod "16e3527e-df72-474b-9905-e770818caa33" (UID: "16e3527e-df72-474b-9905-e770818caa33"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 05:54:46 crc kubenswrapper[4799]: I1129 05:54:46.459249 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rd744\" (UniqueName: \"kubernetes.io/projected/16e3527e-df72-474b-9905-e770818caa33-kube-api-access-rd744\") pod \"16e3527e-df72-474b-9905-e770818caa33\" (UID: \"16e3527e-df72-474b-9905-e770818caa33\") " Nov 29 05:54:46 crc kubenswrapper[4799]: I1129 05:54:46.460153 4799 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/16e3527e-df72-474b-9905-e770818caa33-host\") on node \"crc\" DevicePath \"\"" Nov 29 05:54:46 crc kubenswrapper[4799]: I1129 05:54:46.922240 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16e3527e-df72-474b-9905-e770818caa33-kube-api-access-rd744" (OuterVolumeSpecName: "kube-api-access-rd744") pod "16e3527e-df72-474b-9905-e770818caa33" (UID: "16e3527e-df72-474b-9905-e770818caa33"). InnerVolumeSpecName "kube-api-access-rd744". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:54:46 crc kubenswrapper[4799]: I1129 05:54:46.991735 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rd744\" (UniqueName: \"kubernetes.io/projected/16e3527e-df72-474b-9905-e770818caa33-kube-api-access-rd744\") on node \"crc\" DevicePath \"\"" Nov 29 05:54:47 crc kubenswrapper[4799]: I1129 05:54:47.226980 4799 scope.go:117] "RemoveContainer" containerID="ec98bac390f9126125f67ab591d022819d6a077ec906b1149d8ccbb45f9c7dd8" Nov 29 05:54:47 crc kubenswrapper[4799]: I1129 05:54:47.227118 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t9ms2/crc-debug-t7hcg" Nov 29 05:54:48 crc kubenswrapper[4799]: I1129 05:54:48.670490 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16e3527e-df72-474b-9905-e770818caa33" path="/var/lib/kubelet/pods/16e3527e-df72-474b-9905-e770818caa33/volumes" Nov 29 05:54:56 crc kubenswrapper[4799]: I1129 05:54:56.659887 4799 scope.go:117] "RemoveContainer" containerID="b560e06b6de1edd9a7f603575cb604d0893a8a24eda3e04d2446b813f4e889f8" Nov 29 05:54:56 crc kubenswrapper[4799]: E1129 05:54:56.660758 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:55:08 crc kubenswrapper[4799]: I1129 05:55:08.659295 4799 scope.go:117] "RemoveContainer" containerID="b560e06b6de1edd9a7f603575cb604d0893a8a24eda3e04d2446b813f4e889f8" Nov 29 05:55:08 crc kubenswrapper[4799]: E1129 05:55:08.660279 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:55:19 crc kubenswrapper[4799]: I1129 05:55:19.659384 4799 scope.go:117] "RemoveContainer" containerID="b560e06b6de1edd9a7f603575cb604d0893a8a24eda3e04d2446b813f4e889f8" Nov 29 05:55:19 crc kubenswrapper[4799]: E1129 05:55:19.660085 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:55:27 crc kubenswrapper[4799]: I1129 05:55:27.435095 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-fdfbddb56-lhh8r_ddd039b3-89fb-4515-a633-e4c8c23bc25c/barbican-api/0.log" Nov 29 05:55:27 crc kubenswrapper[4799]: I1129 05:55:27.648020 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-fdfbddb56-lhh8r_ddd039b3-89fb-4515-a633-e4c8c23bc25c/barbican-api-log/0.log" Nov 29 05:55:27 crc kubenswrapper[4799]: I1129 05:55:27.706523 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6495665bf6-xn8f7_bdd214a8-9094-42ef-b2c6-2452b59596f7/barbican-keystone-listener/0.log" Nov 29 05:55:27 crc kubenswrapper[4799]: I1129 05:55:27.913703 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-76859cbf9f-kdxtl_f14d7038-795e-4f8a-8ce7-81e64c86137b/barbican-worker/0.log" Nov 29 05:55:27 crc kubenswrapper[4799]: I1129 05:55:27.920274 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-76859cbf9f-kdxtl_f14d7038-795e-4f8a-8ce7-81e64c86137b/barbican-worker-log/0.log" Nov 29 05:55:27 crc kubenswrapper[4799]: I1129 05:55:27.927984 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6495665bf6-xn8f7_bdd214a8-9094-42ef-b2c6-2452b59596f7/barbican-keystone-listener-log/0.log" Nov 29 05:55:28 crc kubenswrapper[4799]: I1129 05:55:28.111300 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-k9rvx_a30efa1d-c61a-4bc1-9350-fa1059e7bb71/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 05:55:28 crc kubenswrapper[4799]: I1129 05:55:28.187363 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c/ceilometer-central-agent/0.log" Nov 29 05:55:28 crc kubenswrapper[4799]: I1129 05:55:28.490810 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c/ceilometer-notification-agent/0.log" Nov 29 05:55:28 crc kubenswrapper[4799]: I1129 05:55:28.521090 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c/sg-core/0.log" Nov 29 05:55:28 crc kubenswrapper[4799]: I1129 05:55:28.567204 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3fadcfdb-01c3-4f90-bac4-9c30a3da6c2c/proxy-httpd/0.log" Nov 29 05:55:28 crc kubenswrapper[4799]: I1129 05:55:28.653109 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-edpm-deployment-openstack-edpm-ipam-fdll7_41294840-0c22-4c85-ae8f-31f3a00458d4/ceph-client-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 05:55:28 crc kubenswrapper[4799]: I1129 05:55:28.842408 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6rm8k_92efc83b-d2e4-44f7-8ede-db45686a6c0f/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 05:55:28 crc kubenswrapper[4799]: I1129 05:55:28.903330 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_64042cd5-9bc5-4371-b063-753befa911c2/cinder-api/0.log" Nov 29 05:55:28 crc kubenswrapper[4799]: I1129 05:55:28.993260 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_64042cd5-9bc5-4371-b063-753befa911c2/cinder-api-log/0.log" Nov 29 05:55:29 crc kubenswrapper[4799]: I1129 05:55:29.139682 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_caa349ef-0de7-46c4-a7cd-fdf43aa6f949/probe/0.log" Nov 29 05:55:29 crc kubenswrapper[4799]: I1129 05:55:29.312633 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_caa349ef-0de7-46c4-a7cd-fdf43aa6f949/cinder-backup/0.log" Nov 29 05:55:29 crc kubenswrapper[4799]: I1129 05:55:29.343155 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_6c7456e1-235c-4bb3-a587-f7e3c620749e/cinder-scheduler/0.log" Nov 29 05:55:29 crc kubenswrapper[4799]: I1129 05:55:29.381484 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_6c7456e1-235c-4bb3-a587-f7e3c620749e/probe/0.log" Nov 29 05:55:29 crc kubenswrapper[4799]: I1129 05:55:29.578573 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_0907967b-0bf5-4fe0-9862-ffbd54efde7f/cinder-volume/0.log" Nov 29 05:55:29 crc kubenswrapper[4799]: I1129 05:55:29.610908 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_0907967b-0bf5-4fe0-9862-ffbd54efde7f/probe/0.log" Nov 29 05:55:29 crc kubenswrapper[4799]: I1129 05:55:29.730993 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-x4w52_86cf64fe-3c75-4b05-8503-a4e3f3e0395c/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 05:55:29 crc kubenswrapper[4799]: I1129 05:55:29.850471 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-k7vvz_56293cbb-ed95-4541-b322-8e86233a8ff5/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 05:55:29 crc kubenswrapper[4799]: I1129 05:55:29.935958 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-76b5fdb995-cbg8h_87fc92de-e3e3-454f-886a-226aa591c580/init/0.log" Nov 29 05:55:30 crc kubenswrapper[4799]: I1129 05:55:30.162291 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-76b5fdb995-cbg8h_87fc92de-e3e3-454f-886a-226aa591c580/init/0.log" Nov 29 05:55:30 crc kubenswrapper[4799]: I1129 05:55:30.180024 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-76b5fdb995-cbg8h_87fc92de-e3e3-454f-886a-226aa591c580/dnsmasq-dns/0.log" Nov 29 05:55:30 crc kubenswrapper[4799]: I1129 05:55:30.182303 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_cd7d046d-4862-47c2-b4d3-e4d81d79706a/glance-httpd/0.log" Nov 29 05:55:30 crc kubenswrapper[4799]: I1129 05:55:30.359887 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_cd7d046d-4862-47c2-b4d3-e4d81d79706a/glance-log/0.log" Nov 29 05:55:30 crc kubenswrapper[4799]: I1129 05:55:30.428056 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_477f0d92-5648-405b-8034-f851200268a9/glance-log/0.log" Nov 29 05:55:30 crc kubenswrapper[4799]: I1129 05:55:30.435415 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_477f0d92-5648-405b-8034-f851200268a9/glance-httpd/0.log" Nov 29 05:55:30 crc kubenswrapper[4799]: I1129 05:55:30.659575 4799 scope.go:117] "RemoveContainer" containerID="b560e06b6de1edd9a7f603575cb604d0893a8a24eda3e04d2446b813f4e889f8" Nov 29 05:55:30 crc kubenswrapper[4799]: E1129 05:55:30.659879 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 05:55:30 crc kubenswrapper[4799]: I1129 05:55:30.710567 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7c5488f496-jvw7p_93fb8bea-4246-40da-b480-c980b628e2b9/horizon/0.log" Nov 29 05:55:30 crc kubenswrapper[4799]: I1129 05:55:30.864865 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7c5488f496-jvw7p_93fb8bea-4246-40da-b480-c980b628e2b9/horizon-log/0.log" Nov 29 05:55:31 crc kubenswrapper[4799]: I1129 05:55:31.272932 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-xcfdv_cbad88f8-b661-4b9f-8815-6c0f2df63d7f/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 05:55:31 crc kubenswrapper[4799]: I1129 05:55:31.319963 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-hdkhl_1784d248-9169-4b88-b8d1-16412106a8dc/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 05:55:31 crc kubenswrapper[4799]: I1129 05:55:31.558075 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29406541-dvcxf_4b4eb06c-ec4a-45f3-b316-34a4410ea46b/keystone-cron/0.log" Nov 29 05:55:31 crc kubenswrapper[4799]: I1129 05:55:31.640675 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_66c61a7f-a655-46e9-8a85-de7ad52bfe6f/kube-state-metrics/0.log" Nov 29 05:55:31 crc kubenswrapper[4799]: I1129 05:55:31.839752 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-v72s9_168cfe40-080e-44fa-863e-8eb8b5b5923c/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 05:55:32 crc kubenswrapper[4799]: I1129 05:55:32.071347 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-757bb6ddd6-9wg88_5d5a9855-0646-45ff-8036-afba2a0d1a41/keystone-api/0.log" Nov 29 05:55:32 crc kubenswrapper[4799]: I1129 05:55:32.259146 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_ef775c01-8e04-4efe-8192-889ddf0ccbd3/probe/0.log" Nov 29 05:55:32 crc kubenswrapper[4799]: I1129 05:55:32.384192 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_ef775c01-8e04-4efe-8192-889ddf0ccbd3/manila-scheduler/0.log" Nov 29 05:55:32 crc kubenswrapper[4799]: I1129 05:55:32.410526 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_b8b16394-158d-47e0-b406-18636e57be57/manila-api/0.log" Nov 29 05:55:32 crc kubenswrapper[4799]: I1129 05:55:32.693298 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_b8b16394-158d-47e0-b406-18636e57be57/manila-api-log/0.log" Nov 29 05:55:32 crc kubenswrapper[4799]: I1129 05:55:32.809715 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_d858c44e-0d42-4c54-b820-dc89173d67d1/probe/0.log" Nov 29 05:55:33 crc kubenswrapper[4799]: I1129 05:55:33.027062 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_d858c44e-0d42-4c54-b820-dc89173d67d1/manila-share/0.log" Nov 29 05:55:33 crc kubenswrapper[4799]: I1129 05:55:33.138226 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-669f887b5-74klb_122be3fa-bc40-45f0-808d-fece3e876c43/neutron-httpd/0.log" Nov 29 05:55:33 crc kubenswrapper[4799]: I1129 05:55:33.206311 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-669f887b5-74klb_122be3fa-bc40-45f0-808d-fece3e876c43/neutron-api/0.log" Nov 29 05:55:33 crc kubenswrapper[4799]: I1129 05:55:33.310468 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjld9_8f234eaf-4d04-4d43-ab4b-aba8a4ab1e34/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 05:55:33 crc kubenswrapper[4799]: I1129 05:55:33.600365 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_cd1dc76e-76ba-4ab4-b40c-587ffbf9b5df/nova-api-log/0.log" Nov 29 05:55:33 crc kubenswrapper[4799]: I1129 05:55:33.862779 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_f6a2c6d8-ac5d-4b14-a907-d624f22d5f1b/nova-cell0-conductor-conductor/0.log" Nov 29 05:55:34 crc kubenswrapper[4799]: I1129 05:55:34.030663 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_27b8e913-bbd1-4a53-ab29-9b31ab8c64c8/nova-cell1-conductor-conductor/0.log" Nov 29 05:55:34 crc kubenswrapper[4799]: I1129 05:55:34.111077 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_cd1dc76e-76ba-4ab4-b40c-587ffbf9b5df/nova-api-api/0.log" Nov 29 05:55:34 crc kubenswrapper[4799]: I1129 05:55:34.188781 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_91d09977-e990-4df4-921c-6f087e18b85f/nova-cell1-novncproxy-novncproxy/0.log" Nov 29 05:55:34 crc kubenswrapper[4799]: I1129 05:55:34.383834 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-4np5b_26a0b212-c652-4475-8210-4b62b653ed79/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 05:55:34 crc kubenswrapper[4799]: I1129 05:55:34.439080 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_bfbbd39d-1e8f-4bdd-991e-71e206bb4ca0/nova-metadata-log/0.log" Nov 29 05:55:34 crc kubenswrapper[4799]: I1129 05:55:34.789020 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_923b93cd-9a7f-4924-8d46-a1cd33612af5/nova-scheduler-scheduler/0.log" Nov 29 05:55:34 crc kubenswrapper[4799]: I1129 05:55:34.815760 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_735902f6-652e-4c85-818d-10f7d9529788/mysql-bootstrap/0.log" Nov 29 05:55:34 crc kubenswrapper[4799]: I1129 05:55:34.961262 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_735902f6-652e-4c85-818d-10f7d9529788/galera/0.log" Nov 29 05:55:35 crc kubenswrapper[4799]: I1129 05:55:35.019904 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_735902f6-652e-4c85-818d-10f7d9529788/mysql-bootstrap/0.log" Nov 29 05:55:35 crc kubenswrapper[4799]: I1129 05:55:35.155620 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_bbc1f3a5-585d-4e6d-a87e-cf77a7b21147/mysql-bootstrap/0.log" Nov 29 05:55:35 crc kubenswrapper[4799]: I1129 05:55:35.343770 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_bbc1f3a5-585d-4e6d-a87e-cf77a7b21147/mysql-bootstrap/0.log" Nov 29 05:55:35 crc kubenswrapper[4799]: I1129 05:55:35.349579 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_bbc1f3a5-585d-4e6d-a87e-cf77a7b21147/galera/0.log" Nov 29 05:55:35 crc kubenswrapper[4799]: I1129 05:55:35.516275 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_548ff572-b2d0-45ec-a131-0ba87044b4ff/openstackclient/0.log" Nov 29 05:55:35 crc kubenswrapper[4799]: I1129 05:55:35.572575 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-kkdw8_1d3d81e5-591a-469d-9851-86271162d455/ovn-controller/0.log" Nov 29 05:55:35 crc kubenswrapper[4799]: I1129 05:55:35.790665 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-5szsr_b0d261b9-8134-46ec-b510-63fce197ccff/openstack-network-exporter/0.log" Nov 29 05:55:36 crc kubenswrapper[4799]: I1129 05:55:36.023555 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-bkkst_e358b742-108b-4f60-989f-2e8eddce487e/ovsdb-server-init/0.log" Nov 29 05:55:36 crc kubenswrapper[4799]: I1129 05:55:36.135935 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_bfbbd39d-1e8f-4bdd-991e-71e206bb4ca0/nova-metadata-metadata/0.log" Nov 29 05:55:36 crc kubenswrapper[4799]: I1129 05:55:36.202910 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-bkkst_e358b742-108b-4f60-989f-2e8eddce487e/ovs-vswitchd/0.log" Nov 29 05:55:36 crc kubenswrapper[4799]: I1129 05:55:36.209075 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-bkkst_e358b742-108b-4f60-989f-2e8eddce487e/ovsdb-server-init/0.log" Nov 29 05:55:36 crc kubenswrapper[4799]: I1129 05:55:36.262984 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-bkkst_e358b742-108b-4f60-989f-2e8eddce487e/ovsdb-server/0.log" Nov 29 05:55:36 crc kubenswrapper[4799]: I1129 05:55:36.463074 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-hnnmv_a9783465-b33b-428e-9691-bc9b341e37ad/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 05:55:36 crc kubenswrapper[4799]: I1129 05:55:36.496494 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_c8c0babb-3163-4a72-a6d5-4fc5e64bee31/openstack-network-exporter/0.log" Nov 29 05:55:36 crc kubenswrapper[4799]: I1129 05:55:36.673505 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_c8c0babb-3163-4a72-a6d5-4fc5e64bee31/ovn-northd/0.log" Nov 29 05:55:36 crc kubenswrapper[4799]: I1129 05:55:36.702322 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_c79e4fc0-8786-40a6-a775-29a4767d072e/ovsdbserver-nb/0.log" Nov 29 05:55:36 crc kubenswrapper[4799]: I1129 05:55:36.737757 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_c79e4fc0-8786-40a6-a775-29a4767d072e/openstack-network-exporter/0.log" Nov 29 05:55:36 crc kubenswrapper[4799]: I1129 05:55:36.891943 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_5bb0e45f-f281-49ec-8a11-37e1ace553db/openstack-network-exporter/0.log" Nov 29 05:55:36 crc kubenswrapper[4799]: I1129 05:55:36.955592 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_5bb0e45f-f281-49ec-8a11-37e1ace553db/ovsdbserver-sb/0.log" Nov 29 05:55:37 crc kubenswrapper[4799]: I1129 05:55:37.147426 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-dc694ffdd-sm2s5_8af48d85-03bf-4d5f-a677-49bc96f0dfee/placement-api/0.log" Nov 29 05:55:37 crc kubenswrapper[4799]: I1129 05:55:37.162488 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac/setup-container/0.log" Nov 29 05:55:37 crc kubenswrapper[4799]: I1129 05:55:37.215526 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-dc694ffdd-sm2s5_8af48d85-03bf-4d5f-a677-49bc96f0dfee/placement-log/0.log" Nov 29 05:55:37 crc kubenswrapper[4799]: I1129 05:55:37.481203 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac/setup-container/0.log" Nov 29 05:55:37 crc kubenswrapper[4799]: I1129 05:55:37.485906 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_1f5c1a3b-7a4e-48df-aaff-de72acf9b1ac/rabbitmq/0.log" Nov 29 05:55:37 crc kubenswrapper[4799]: I1129 05:55:37.508756 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_0bfecf67-24fe-42fa-bad5-4a597ad9d3af/setup-container/0.log" Nov 29 05:55:37 crc kubenswrapper[4799]: I1129 05:55:37.667764 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_0bfecf67-24fe-42fa-bad5-4a597ad9d3af/setup-container/0.log" Nov 29 05:55:37 crc kubenswrapper[4799]: I1129 05:55:37.748940 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_0bfecf67-24fe-42fa-bad5-4a597ad9d3af/rabbitmq/0.log" Nov 29 05:55:37 crc kubenswrapper[4799]: I1129 05:55:37.793512 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-d68wr_0a726ce8-7d0d-4906-a2a4-e6f146dc4635/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 05:55:37 crc kubenswrapper[4799]: I1129 05:55:37.920181 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-vj9bq_3b36b813-5af5-4cc9-92c9-818aa2b99423/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 05:55:38 crc kubenswrapper[4799]: I1129 05:55:38.079497 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-h4p62_828374a6-f506-4e41-83fe-32bf0f6a4ed3/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 05:55:38 crc kubenswrapper[4799]: I1129 05:55:38.172900 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-9hpsr_3a84e2da-773c-4bf7-8f99-a84276791e78/ssh-known-hosts-edpm-deployment/0.log" Nov 29 05:55:38 crc kubenswrapper[4799]: I1129 05:55:38.360563 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_c4a30984-7c0c-493b-b752-b603232c837f/tempest-tests-tempest-tests-runner/0.log" Nov 29 05:55:38 crc kubenswrapper[4799]: I1129 05:55:38.450747 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_ec2916e5-c59e-4909-9c3c-da11e14e2413/test-operator-logs-container/0.log" Nov 29 05:55:38 crc kubenswrapper[4799]: I1129 05:55:38.566074 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-gnq8m_91c293ad-ab01-40e7-9fb6-b8ef99152e62/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 05:55:42 crc kubenswrapper[4799]: I1129 05:55:42.658152 4799 scope.go:117] "RemoveContainer" containerID="b560e06b6de1edd9a7f603575cb604d0893a8a24eda3e04d2446b813f4e889f8" Nov 29 05:55:43 crc kubenswrapper[4799]: I1129 05:55:43.767530 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" event={"ID":"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8","Type":"ContainerStarted","Data":"53954acbaab2bd75534d88a03d4525340098eb9e1108f075c344f3be17a8ee5e"} Nov 29 05:55:58 crc kubenswrapper[4799]: I1129 05:55:58.770001 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_b6f5ef6d-bec8-4b82-8ebc-53952954134c/memcached/0.log" Nov 29 05:56:07 crc kubenswrapper[4799]: I1129 05:56:07.317332 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_232b41aac13e001dc857b9795d2b854afae448fcc1728af84c1ef8a484ctr76_e3b7bcee-1d1a-41a2-b1ea-221fe74998ef/util/0.log" Nov 29 05:56:07 crc kubenswrapper[4799]: I1129 05:56:07.518430 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_232b41aac13e001dc857b9795d2b854afae448fcc1728af84c1ef8a484ctr76_e3b7bcee-1d1a-41a2-b1ea-221fe74998ef/pull/0.log" Nov 29 05:56:07 crc kubenswrapper[4799]: I1129 05:56:07.528427 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_232b41aac13e001dc857b9795d2b854afae448fcc1728af84c1ef8a484ctr76_e3b7bcee-1d1a-41a2-b1ea-221fe74998ef/pull/0.log" Nov 29 05:56:07 crc kubenswrapper[4799]: I1129 05:56:07.558362 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_232b41aac13e001dc857b9795d2b854afae448fcc1728af84c1ef8a484ctr76_e3b7bcee-1d1a-41a2-b1ea-221fe74998ef/util/0.log" Nov 29 05:56:07 crc kubenswrapper[4799]: I1129 05:56:07.697113 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_232b41aac13e001dc857b9795d2b854afae448fcc1728af84c1ef8a484ctr76_e3b7bcee-1d1a-41a2-b1ea-221fe74998ef/util/0.log" Nov 29 05:56:07 crc kubenswrapper[4799]: I1129 05:56:07.740459 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_232b41aac13e001dc857b9795d2b854afae448fcc1728af84c1ef8a484ctr76_e3b7bcee-1d1a-41a2-b1ea-221fe74998ef/extract/0.log" Nov 29 05:56:07 crc kubenswrapper[4799]: I1129 05:56:07.741527 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_232b41aac13e001dc857b9795d2b854afae448fcc1728af84c1ef8a484ctr76_e3b7bcee-1d1a-41a2-b1ea-221fe74998ef/pull/0.log" Nov 29 05:56:07 crc kubenswrapper[4799]: I1129 05:56:07.873255 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-6vnhr_7232eadb-c887-440b-b6f9-bffa4a6dc871/kube-rbac-proxy/0.log" Nov 29 05:56:07 crc kubenswrapper[4799]: I1129 05:56:07.922258 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-7zddj_259abc1e-a30c-4a8b-acd1-18052862198e/kube-rbac-proxy/0.log" Nov 29 05:56:07 crc kubenswrapper[4799]: I1129 05:56:07.958665 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-6vnhr_7232eadb-c887-440b-b6f9-bffa4a6dc871/manager/0.log" Nov 29 05:56:08 crc kubenswrapper[4799]: I1129 05:56:08.082693 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-7zddj_259abc1e-a30c-4a8b-acd1-18052862198e/manager/0.log" Nov 29 05:56:08 crc kubenswrapper[4799]: I1129 05:56:08.177391 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-pkrhh_89038368-29a3-417e-8ffa-819e5901a52a/manager/0.log" Nov 29 05:56:08 crc kubenswrapper[4799]: I1129 05:56:08.191288 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-pkrhh_89038368-29a3-417e-8ffa-819e5901a52a/kube-rbac-proxy/0.log" Nov 29 05:56:08 crc kubenswrapper[4799]: I1129 05:56:08.314178 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-668d9c48b9-lkbfl_743ba96f-315f-4e0d-abb0-5a6389140c7b/kube-rbac-proxy/0.log" Nov 29 05:56:08 crc kubenswrapper[4799]: I1129 05:56:08.433498 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-668d9c48b9-lkbfl_743ba96f-315f-4e0d-abb0-5a6389140c7b/manager/0.log" Nov 29 05:56:08 crc kubenswrapper[4799]: I1129 05:56:08.497957 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-p555g_2798aeab-f9b9-4728-ad63-186b16afcc31/kube-rbac-proxy/0.log" Nov 29 05:56:08 crc kubenswrapper[4799]: I1129 05:56:08.563410 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-p555g_2798aeab-f9b9-4728-ad63-186b16afcc31/manager/0.log" Nov 29 05:56:08 crc kubenswrapper[4799]: I1129 05:56:08.632051 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-kxgw8_94ec7bc8-8413-4040-90dd-a5c9ce4ca98f/kube-rbac-proxy/0.log" Nov 29 05:56:08 crc kubenswrapper[4799]: I1129 05:56:08.705279 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-kxgw8_94ec7bc8-8413-4040-90dd-a5c9ce4ca98f/manager/0.log" Nov 29 05:56:08 crc kubenswrapper[4799]: I1129 05:56:08.794894 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-pn7wx_8c9ec062-c2b9-47a6-8d7a-eb9f00e63ef3/kube-rbac-proxy/0.log" Nov 29 05:56:08 crc kubenswrapper[4799]: I1129 05:56:08.957860 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-pn7wx_8c9ec062-c2b9-47a6-8d7a-eb9f00e63ef3/manager/0.log" Nov 29 05:56:08 crc kubenswrapper[4799]: I1129 05:56:08.983908 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-25fzb_ff72829c-dfd8-4ae2-b508-f89e40f654eb/kube-rbac-proxy/0.log" Nov 29 05:56:09 crc kubenswrapper[4799]: I1129 05:56:09.000385 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-25fzb_ff72829c-dfd8-4ae2-b508-f89e40f654eb/manager/0.log" Nov 29 05:56:09 crc kubenswrapper[4799]: I1129 05:56:09.139120 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-546d4bdf48-8p2z8_f696db58-bd41-4db0-9ec7-c4a4710e6b8e/kube-rbac-proxy/0.log" Nov 29 05:56:09 crc kubenswrapper[4799]: I1129 05:56:09.187754 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-546d4bdf48-8p2z8_f696db58-bd41-4db0-9ec7-c4a4710e6b8e/manager/0.log" Nov 29 05:56:09 crc kubenswrapper[4799]: I1129 05:56:09.281024 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-84f754f96-t5gjl_4bf2a041-0f89-480e-88d2-ba630db189c1/kube-rbac-proxy/0.log" Nov 29 05:56:09 crc kubenswrapper[4799]: I1129 05:56:09.378120 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-84f754f96-t5gjl_4bf2a041-0f89-480e-88d2-ba630db189c1/manager/0.log" Nov 29 05:56:09 crc kubenswrapper[4799]: I1129 05:56:09.476750 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-sfpmg_7851df67-88d5-45d8-804e-5d3bf3fdb3af/kube-rbac-proxy/0.log" Nov 29 05:56:09 crc kubenswrapper[4799]: I1129 05:56:09.478671 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-sfpmg_7851df67-88d5-45d8-804e-5d3bf3fdb3af/manager/0.log" Nov 29 05:56:09 crc kubenswrapper[4799]: I1129 05:56:09.572315 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-pc2zl_e1d949de-4ce7-4016-83fd-4c0ce003eb0f/kube-rbac-proxy/0.log" Nov 29 05:56:09 crc kubenswrapper[4799]: I1129 05:56:09.691895 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-pc2zl_e1d949de-4ce7-4016-83fd-4c0ce003eb0f/manager/0.log" Nov 29 05:56:09 crc kubenswrapper[4799]: I1129 05:56:09.784532 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-8nl87_bdcce387-cd82-490e-bd50-cf47158b1ea6/kube-rbac-proxy/0.log" Nov 29 05:56:09 crc kubenswrapper[4799]: I1129 05:56:09.873384 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-8nl87_bdcce387-cd82-490e-bd50-cf47158b1ea6/manager/0.log" Nov 29 05:56:09 crc kubenswrapper[4799]: I1129 05:56:09.925410 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-wstbs_bbe2e2fb-4619-405c-a9ad-4bdf727882cd/kube-rbac-proxy/0.log" Nov 29 05:56:09 crc kubenswrapper[4799]: I1129 05:56:09.991302 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-wstbs_bbe2e2fb-4619-405c-a9ad-4bdf727882cd/manager/0.log" Nov 29 05:56:10 crc kubenswrapper[4799]: I1129 05:56:10.781968 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4qnwnt_8e0677fe-7f53-4e50-881b-943fffe1c0ff/kube-rbac-proxy/0.log" Nov 29 05:56:10 crc kubenswrapper[4799]: I1129 05:56:10.883622 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4qnwnt_8e0677fe-7f53-4e50-881b-943fffe1c0ff/manager/0.log" Nov 29 05:56:11 crc kubenswrapper[4799]: I1129 05:56:11.112385 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-hr4n9_8a6a8fd6-7ba3-4457-9d23-b18414836f30/registry-server/0.log" Nov 29 05:56:11 crc kubenswrapper[4799]: I1129 05:56:11.270294 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-5bb85d56cf-bx84s_70a5864d-82f9-416d-a713-8e1bfa00733d/operator/0.log" Nov 29 05:56:11 crc kubenswrapper[4799]: I1129 05:56:11.278094 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-qfp99_888d5a2b-8a13-4042-8c43-59241cfdc088/kube-rbac-proxy/0.log" Nov 29 05:56:11 crc kubenswrapper[4799]: I1129 05:56:11.415432 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-qfp99_888d5a2b-8a13-4042-8c43-59241cfdc088/manager/0.log" Nov 29 05:56:11 crc kubenswrapper[4799]: I1129 05:56:11.490741 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-4p69t_861aea23-001f-4259-b69a-64b95db49c73/kube-rbac-proxy/0.log" Nov 29 05:56:11 crc kubenswrapper[4799]: I1129 05:56:11.553266 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-4p69t_861aea23-001f-4259-b69a-64b95db49c73/manager/0.log" Nov 29 05:56:11 crc kubenswrapper[4799]: I1129 05:56:11.745671 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-fsl4h_225a8e52-78ae-41eb-9c27-64edb57a135e/operator/0.log" Nov 29 05:56:11 crc kubenswrapper[4799]: I1129 05:56:11.821186 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-4xvn4_ac90f9ba-d2ef-44a1-84e7-8e809b5eef13/kube-rbac-proxy/0.log" Nov 29 05:56:11 crc kubenswrapper[4799]: I1129 05:56:11.864407 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-4xvn4_ac90f9ba-d2ef-44a1-84e7-8e809b5eef13/manager/0.log" Nov 29 05:56:12 crc kubenswrapper[4799]: I1129 05:56:12.079852 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-576b8587b8-cftqf_41eeb1eb-9b97-41fa-ad22-09d4d4b34abc/manager/0.log" Nov 29 05:56:12 crc kubenswrapper[4799]: I1129 05:56:12.252729 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-wv65n_06627027-315a-482d-9e26-32c5b4df115a/kube-rbac-proxy/0.log" Nov 29 05:56:12 crc kubenswrapper[4799]: I1129 05:56:12.282338 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-tbth4_1bb0507a-76c1-4a53-971c-dce960b71242/kube-rbac-proxy/0.log" Nov 29 05:56:12 crc kubenswrapper[4799]: I1129 05:56:12.321453 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-wv65n_06627027-315a-482d-9e26-32c5b4df115a/manager/0.log" Nov 29 05:56:12 crc kubenswrapper[4799]: I1129 05:56:12.478105 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-v4kv7_1edee957-6d50-48c2-bccc-e6702bcb5ce0/kube-rbac-proxy/0.log" Nov 29 05:56:12 crc kubenswrapper[4799]: I1129 05:56:12.497901 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-tbth4_1bb0507a-76c1-4a53-971c-dce960b71242/manager/0.log" Nov 29 05:56:12 crc kubenswrapper[4799]: I1129 05:56:12.529415 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-v4kv7_1edee957-6d50-48c2-bccc-e6702bcb5ce0/manager/0.log" Nov 29 05:56:31 crc kubenswrapper[4799]: I1129 05:56:31.194066 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-85llw_bfdd16e4-1cb8-4178-af49-4db764abc507/control-plane-machine-set-operator/0.log" Nov 29 05:56:31 crc kubenswrapper[4799]: I1129 05:56:31.883999 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-wznnv_daa3f105-2960-45a0-ab05-cb2cedea7e9f/kube-rbac-proxy/0.log" Nov 29 05:56:31 crc kubenswrapper[4799]: I1129 05:56:31.885402 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-wznnv_daa3f105-2960-45a0-ab05-cb2cedea7e9f/machine-api-operator/0.log" Nov 29 05:56:44 crc kubenswrapper[4799]: I1129 05:56:44.081417 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-4hw5r_9347b23b-a5d2-40f4-8d55-a320f5a0c78a/cert-manager-controller/0.log" Nov 29 05:56:44 crc kubenswrapper[4799]: I1129 05:56:44.244753 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-68jsj_90513767-7b60-461e-bb69-727a8f25e15e/cert-manager-cainjector/0.log" Nov 29 05:56:44 crc kubenswrapper[4799]: I1129 05:56:44.315391 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-b9lmh_28341ac5-a271-4435-a855-512a548beb96/cert-manager-webhook/0.log" Nov 29 05:56:57 crc kubenswrapper[4799]: I1129 05:56:57.448956 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-j5cvx_bb0e8eeb-0667-40e9-b814-33a66c0e78a0/nmstate-console-plugin/0.log" Nov 29 05:56:57 crc kubenswrapper[4799]: I1129 05:56:57.644169 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-7k5v8_767c1d25-4a7d-49b3-8180-8f46793649be/nmstate-handler/0.log" Nov 29 05:56:57 crc kubenswrapper[4799]: I1129 05:56:57.658047 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-2955x_e96ac85b-718a-45dc-be96-b514bffa1431/kube-rbac-proxy/0.log" Nov 29 05:56:57 crc kubenswrapper[4799]: I1129 05:56:57.705117 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-2955x_e96ac85b-718a-45dc-be96-b514bffa1431/nmstate-metrics/0.log" Nov 29 05:56:57 crc kubenswrapper[4799]: I1129 05:56:57.865657 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-7wvt6_9924770d-6e30-4da4-959a-40b3eade5bf3/nmstate-operator/0.log" Nov 29 05:56:57 crc kubenswrapper[4799]: I1129 05:56:57.930067 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-2kxrf_337961b2-38e7-4e32-831a-0c970535fecd/nmstate-webhook/0.log" Nov 29 05:57:12 crc kubenswrapper[4799]: I1129 05:57:12.478455 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-27dxl_9b8b8dfd-90f2-44b4-921a-e56563c76655/kube-rbac-proxy/0.log" Nov 29 05:57:12 crc kubenswrapper[4799]: I1129 05:57:12.573477 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-27dxl_9b8b8dfd-90f2-44b4-921a-e56563c76655/controller/0.log" Nov 29 05:57:12 crc kubenswrapper[4799]: I1129 05:57:12.687851 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95pk_bae7578a-afd9-462d-b824-beae428d9edf/cp-frr-files/0.log" Nov 29 05:57:12 crc kubenswrapper[4799]: I1129 05:57:12.923623 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95pk_bae7578a-afd9-462d-b824-beae428d9edf/cp-frr-files/0.log" Nov 29 05:57:12 crc kubenswrapper[4799]: I1129 05:57:12.933100 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95pk_bae7578a-afd9-462d-b824-beae428d9edf/cp-reloader/0.log" Nov 29 05:57:12 crc kubenswrapper[4799]: I1129 05:57:12.937835 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95pk_bae7578a-afd9-462d-b824-beae428d9edf/cp-reloader/0.log" Nov 29 05:57:12 crc kubenswrapper[4799]: I1129 05:57:12.950572 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95pk_bae7578a-afd9-462d-b824-beae428d9edf/cp-metrics/0.log" Nov 29 05:57:13 crc kubenswrapper[4799]: I1129 05:57:13.101637 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95pk_bae7578a-afd9-462d-b824-beae428d9edf/cp-frr-files/0.log" Nov 29 05:57:13 crc kubenswrapper[4799]: I1129 05:57:13.124998 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95pk_bae7578a-afd9-462d-b824-beae428d9edf/cp-metrics/0.log" Nov 29 05:57:13 crc kubenswrapper[4799]: I1129 05:57:13.155224 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95pk_bae7578a-afd9-462d-b824-beae428d9edf/cp-metrics/0.log" Nov 29 05:57:13 crc kubenswrapper[4799]: I1129 05:57:13.196972 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95pk_bae7578a-afd9-462d-b824-beae428d9edf/cp-reloader/0.log" Nov 29 05:57:13 crc kubenswrapper[4799]: I1129 05:57:13.303225 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95pk_bae7578a-afd9-462d-b824-beae428d9edf/cp-metrics/0.log" Nov 29 05:57:13 crc kubenswrapper[4799]: I1129 05:57:13.328191 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95pk_bae7578a-afd9-462d-b824-beae428d9edf/cp-reloader/0.log" Nov 29 05:57:13 crc kubenswrapper[4799]: I1129 05:57:13.328602 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95pk_bae7578a-afd9-462d-b824-beae428d9edf/cp-frr-files/0.log" Nov 29 05:57:13 crc kubenswrapper[4799]: I1129 05:57:13.408117 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95pk_bae7578a-afd9-462d-b824-beae428d9edf/controller/0.log" Nov 29 05:57:13 crc kubenswrapper[4799]: I1129 05:57:13.512085 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95pk_bae7578a-afd9-462d-b824-beae428d9edf/frr-metrics/0.log" Nov 29 05:57:13 crc kubenswrapper[4799]: I1129 05:57:13.534020 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95pk_bae7578a-afd9-462d-b824-beae428d9edf/kube-rbac-proxy/0.log" Nov 29 05:57:13 crc kubenswrapper[4799]: I1129 05:57:13.681605 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95pk_bae7578a-afd9-462d-b824-beae428d9edf/kube-rbac-proxy-frr/0.log" Nov 29 05:57:13 crc kubenswrapper[4799]: I1129 05:57:13.712335 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95pk_bae7578a-afd9-462d-b824-beae428d9edf/reloader/0.log" Nov 29 05:57:13 crc kubenswrapper[4799]: I1129 05:57:13.935461 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-j9pvr_95a0cc49-7fb2-4ed4-85a5-50d4248f07aa/frr-k8s-webhook-server/0.log" Nov 29 05:57:13 crc kubenswrapper[4799]: I1129 05:57:13.986214 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-6887d9747b-t425t_cabb2055-e31c-4a75-a9f9-d403684efeec/manager/0.log" Nov 29 05:57:14 crc kubenswrapper[4799]: I1129 05:57:14.190182 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-78b9fb955c-gptcp_dd389a73-2aa5-4f69-8ff1-5a3e7a421c1e/webhook-server/0.log" Nov 29 05:57:14 crc kubenswrapper[4799]: I1129 05:57:14.428734 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-tc7fc_0845b467-e744-408a-b2e7-d61b027724dd/kube-rbac-proxy/0.log" Nov 29 05:57:14 crc kubenswrapper[4799]: I1129 05:57:14.950571 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-tc7fc_0845b467-e744-408a-b2e7-d61b027724dd/speaker/0.log" Nov 29 05:57:15 crc kubenswrapper[4799]: I1129 05:57:15.023301 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95pk_bae7578a-afd9-462d-b824-beae428d9edf/frr/0.log" Nov 29 05:57:27 crc kubenswrapper[4799]: I1129 05:57:27.483278 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftbmdt_01a9ddbf-bda8-4f39-b22e-400f1a1505b0/util/0.log" Nov 29 05:57:28 crc kubenswrapper[4799]: I1129 05:57:28.048501 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftbmdt_01a9ddbf-bda8-4f39-b22e-400f1a1505b0/util/0.log" Nov 29 05:57:28 crc kubenswrapper[4799]: I1129 05:57:28.076171 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftbmdt_01a9ddbf-bda8-4f39-b22e-400f1a1505b0/pull/0.log" Nov 29 05:57:28 crc kubenswrapper[4799]: I1129 05:57:28.078391 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftbmdt_01a9ddbf-bda8-4f39-b22e-400f1a1505b0/pull/0.log" Nov 29 05:57:28 crc kubenswrapper[4799]: I1129 05:57:28.260731 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftbmdt_01a9ddbf-bda8-4f39-b22e-400f1a1505b0/pull/0.log" Nov 29 05:57:28 crc kubenswrapper[4799]: I1129 05:57:28.268030 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftbmdt_01a9ddbf-bda8-4f39-b22e-400f1a1505b0/util/0.log" Nov 29 05:57:28 crc kubenswrapper[4799]: I1129 05:57:28.287846 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftbmdt_01a9ddbf-bda8-4f39-b22e-400f1a1505b0/extract/0.log" Nov 29 05:57:28 crc kubenswrapper[4799]: I1129 05:57:28.413705 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83flptb_c94962e4-e51f-4e30-9c0e-75a19dfa16d9/util/0.log" Nov 29 05:57:28 crc kubenswrapper[4799]: I1129 05:57:28.586992 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83flptb_c94962e4-e51f-4e30-9c0e-75a19dfa16d9/pull/0.log" Nov 29 05:57:28 crc kubenswrapper[4799]: I1129 05:57:28.587166 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83flptb_c94962e4-e51f-4e30-9c0e-75a19dfa16d9/util/0.log" Nov 29 05:57:28 crc kubenswrapper[4799]: I1129 05:57:28.615490 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83flptb_c94962e4-e51f-4e30-9c0e-75a19dfa16d9/pull/0.log" Nov 29 05:57:28 crc kubenswrapper[4799]: I1129 05:57:28.809110 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83flptb_c94962e4-e51f-4e30-9c0e-75a19dfa16d9/util/0.log" Nov 29 05:57:28 crc kubenswrapper[4799]: I1129 05:57:28.809143 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83flptb_c94962e4-e51f-4e30-9c0e-75a19dfa16d9/pull/0.log" Nov 29 05:57:28 crc kubenswrapper[4799]: I1129 05:57:28.822092 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83flptb_c94962e4-e51f-4e30-9c0e-75a19dfa16d9/extract/0.log" Nov 29 05:57:28 crc kubenswrapper[4799]: I1129 05:57:28.984331 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pmw7d_cfabf29e-ad4c-4ded-b754-5b02a25ff34b/extract-utilities/0.log" Nov 29 05:57:29 crc kubenswrapper[4799]: I1129 05:57:29.149650 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pmw7d_cfabf29e-ad4c-4ded-b754-5b02a25ff34b/extract-content/0.log" Nov 29 05:57:29 crc kubenswrapper[4799]: I1129 05:57:29.192239 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pmw7d_cfabf29e-ad4c-4ded-b754-5b02a25ff34b/extract-utilities/0.log" Nov 29 05:57:29 crc kubenswrapper[4799]: I1129 05:57:29.196493 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pmw7d_cfabf29e-ad4c-4ded-b754-5b02a25ff34b/extract-content/0.log" Nov 29 05:57:29 crc kubenswrapper[4799]: I1129 05:57:29.356322 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pmw7d_cfabf29e-ad4c-4ded-b754-5b02a25ff34b/extract-content/0.log" Nov 29 05:57:29 crc kubenswrapper[4799]: I1129 05:57:29.383726 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pmw7d_cfabf29e-ad4c-4ded-b754-5b02a25ff34b/extract-utilities/0.log" Nov 29 05:57:29 crc kubenswrapper[4799]: I1129 05:57:29.610018 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-wc7hk_3642ba6f-6029-462d-9441-903056f9886a/extract-utilities/0.log" Nov 29 05:57:29 crc kubenswrapper[4799]: I1129 05:57:29.880908 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-wc7hk_3642ba6f-6029-462d-9441-903056f9886a/extract-content/0.log" Nov 29 05:57:29 crc kubenswrapper[4799]: I1129 05:57:29.899479 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-wc7hk_3642ba6f-6029-462d-9441-903056f9886a/extract-content/0.log" Nov 29 05:57:29 crc kubenswrapper[4799]: I1129 05:57:29.928519 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-wc7hk_3642ba6f-6029-462d-9441-903056f9886a/extract-utilities/0.log" Nov 29 05:57:30 crc kubenswrapper[4799]: I1129 05:57:30.045628 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pmw7d_cfabf29e-ad4c-4ded-b754-5b02a25ff34b/registry-server/0.log" Nov 29 05:57:30 crc kubenswrapper[4799]: I1129 05:57:30.129994 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-wc7hk_3642ba6f-6029-462d-9441-903056f9886a/extract-utilities/0.log" Nov 29 05:57:30 crc kubenswrapper[4799]: I1129 05:57:30.168775 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-wc7hk_3642ba6f-6029-462d-9441-903056f9886a/extract-content/0.log" Nov 29 05:57:30 crc kubenswrapper[4799]: I1129 05:57:30.270385 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-wc7hk_3642ba6f-6029-462d-9441-903056f9886a/registry-server/0.log" Nov 29 05:57:30 crc kubenswrapper[4799]: I1129 05:57:30.380769 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-szfkm_72063360-3635-4677-8d2a-8d639af3ad6d/marketplace-operator/0.log" Nov 29 05:57:30 crc kubenswrapper[4799]: I1129 05:57:30.453003 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qwxcg_6ee05926-5af3-4fac-9772-8e1484c4de17/extract-utilities/0.log" Nov 29 05:57:30 crc kubenswrapper[4799]: I1129 05:57:30.611746 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qwxcg_6ee05926-5af3-4fac-9772-8e1484c4de17/extract-utilities/0.log" Nov 29 05:57:30 crc kubenswrapper[4799]: I1129 05:57:30.636942 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qwxcg_6ee05926-5af3-4fac-9772-8e1484c4de17/extract-content/0.log" Nov 29 05:57:30 crc kubenswrapper[4799]: I1129 05:57:30.637042 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qwxcg_6ee05926-5af3-4fac-9772-8e1484c4de17/extract-content/0.log" Nov 29 05:57:30 crc kubenswrapper[4799]: I1129 05:57:30.843052 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qwxcg_6ee05926-5af3-4fac-9772-8e1484c4de17/extract-content/0.log" Nov 29 05:57:30 crc kubenswrapper[4799]: I1129 05:57:30.848967 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qwxcg_6ee05926-5af3-4fac-9772-8e1484c4de17/extract-utilities/0.log" Nov 29 05:57:30 crc kubenswrapper[4799]: I1129 05:57:30.995241 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qwxcg_6ee05926-5af3-4fac-9772-8e1484c4de17/registry-server/0.log" Nov 29 05:57:31 crc kubenswrapper[4799]: I1129 05:57:31.049583 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xngqr_47f0c357-1fc0-4646-ae33-b4497ac972e7/extract-utilities/0.log" Nov 29 05:57:31 crc kubenswrapper[4799]: I1129 05:57:31.175309 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xngqr_47f0c357-1fc0-4646-ae33-b4497ac972e7/extract-content/0.log" Nov 29 05:57:31 crc kubenswrapper[4799]: I1129 05:57:31.175329 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xngqr_47f0c357-1fc0-4646-ae33-b4497ac972e7/extract-utilities/0.log" Nov 29 05:57:31 crc kubenswrapper[4799]: I1129 05:57:31.175492 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xngqr_47f0c357-1fc0-4646-ae33-b4497ac972e7/extract-content/0.log" Nov 29 05:57:31 crc kubenswrapper[4799]: I1129 05:57:31.336317 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xngqr_47f0c357-1fc0-4646-ae33-b4497ac972e7/extract-utilities/0.log" Nov 29 05:57:31 crc kubenswrapper[4799]: I1129 05:57:31.347811 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xngqr_47f0c357-1fc0-4646-ae33-b4497ac972e7/extract-content/0.log" Nov 29 05:57:31 crc kubenswrapper[4799]: I1129 05:57:31.957827 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xngqr_47f0c357-1fc0-4646-ae33-b4497ac972e7/registry-server/0.log" Nov 29 05:57:53 crc kubenswrapper[4799]: I1129 05:57:53.051370 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5994b"] Nov 29 05:57:53 crc kubenswrapper[4799]: E1129 05:57:53.053067 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16e3527e-df72-474b-9905-e770818caa33" containerName="container-00" Nov 29 05:57:53 crc kubenswrapper[4799]: I1129 05:57:53.053088 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="16e3527e-df72-474b-9905-e770818caa33" containerName="container-00" Nov 29 05:57:53 crc kubenswrapper[4799]: I1129 05:57:53.053345 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="16e3527e-df72-474b-9905-e770818caa33" containerName="container-00" Nov 29 05:57:53 crc kubenswrapper[4799]: I1129 05:57:53.055235 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5994b" Nov 29 05:57:53 crc kubenswrapper[4799]: I1129 05:57:53.073771 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5994b"] Nov 29 05:57:53 crc kubenswrapper[4799]: I1129 05:57:53.112265 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6da13658-9a5a-44cd-944d-7bc1891c812c-catalog-content\") pod \"community-operators-5994b\" (UID: \"6da13658-9a5a-44cd-944d-7bc1891c812c\") " pod="openshift-marketplace/community-operators-5994b" Nov 29 05:57:53 crc kubenswrapper[4799]: I1129 05:57:53.112387 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52bj6\" (UniqueName: \"kubernetes.io/projected/6da13658-9a5a-44cd-944d-7bc1891c812c-kube-api-access-52bj6\") pod \"community-operators-5994b\" (UID: \"6da13658-9a5a-44cd-944d-7bc1891c812c\") " pod="openshift-marketplace/community-operators-5994b" Nov 29 05:57:53 crc kubenswrapper[4799]: I1129 05:57:53.112482 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6da13658-9a5a-44cd-944d-7bc1891c812c-utilities\") pod \"community-operators-5994b\" (UID: \"6da13658-9a5a-44cd-944d-7bc1891c812c\") " pod="openshift-marketplace/community-operators-5994b" Nov 29 05:57:53 crc kubenswrapper[4799]: I1129 05:57:53.213727 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52bj6\" (UniqueName: \"kubernetes.io/projected/6da13658-9a5a-44cd-944d-7bc1891c812c-kube-api-access-52bj6\") pod \"community-operators-5994b\" (UID: \"6da13658-9a5a-44cd-944d-7bc1891c812c\") " pod="openshift-marketplace/community-operators-5994b" Nov 29 05:57:53 crc kubenswrapper[4799]: I1129 05:57:53.213855 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6da13658-9a5a-44cd-944d-7bc1891c812c-utilities\") pod \"community-operators-5994b\" (UID: \"6da13658-9a5a-44cd-944d-7bc1891c812c\") " pod="openshift-marketplace/community-operators-5994b" Nov 29 05:57:53 crc kubenswrapper[4799]: I1129 05:57:53.213921 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6da13658-9a5a-44cd-944d-7bc1891c812c-catalog-content\") pod \"community-operators-5994b\" (UID: \"6da13658-9a5a-44cd-944d-7bc1891c812c\") " pod="openshift-marketplace/community-operators-5994b" Nov 29 05:57:53 crc kubenswrapper[4799]: I1129 05:57:53.214307 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6da13658-9a5a-44cd-944d-7bc1891c812c-utilities\") pod \"community-operators-5994b\" (UID: \"6da13658-9a5a-44cd-944d-7bc1891c812c\") " pod="openshift-marketplace/community-operators-5994b" Nov 29 05:57:53 crc kubenswrapper[4799]: I1129 05:57:53.214393 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6da13658-9a5a-44cd-944d-7bc1891c812c-catalog-content\") pod \"community-operators-5994b\" (UID: \"6da13658-9a5a-44cd-944d-7bc1891c812c\") " pod="openshift-marketplace/community-operators-5994b" Nov 29 05:57:53 crc kubenswrapper[4799]: I1129 05:57:53.233512 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52bj6\" (UniqueName: \"kubernetes.io/projected/6da13658-9a5a-44cd-944d-7bc1891c812c-kube-api-access-52bj6\") pod \"community-operators-5994b\" (UID: \"6da13658-9a5a-44cd-944d-7bc1891c812c\") " pod="openshift-marketplace/community-operators-5994b" Nov 29 05:57:53 crc kubenswrapper[4799]: I1129 05:57:53.378722 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5994b" Nov 29 05:57:54 crc kubenswrapper[4799]: I1129 05:57:54.000100 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5994b"] Nov 29 05:57:54 crc kubenswrapper[4799]: I1129 05:57:54.873240 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5994b" event={"ID":"6da13658-9a5a-44cd-944d-7bc1891c812c","Type":"ContainerStarted","Data":"f243205ea20722f5af6fc2a958aefd914076ee90a4f0f04c1a0bb2add06aef62"} Nov 29 05:57:54 crc kubenswrapper[4799]: I1129 05:57:54.873603 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5994b" event={"ID":"6da13658-9a5a-44cd-944d-7bc1891c812c","Type":"ContainerStarted","Data":"88f9923712615b0d6d98d7e60a0fc371f9466e69c7877255fc39d15c406b2502"} Nov 29 05:57:55 crc kubenswrapper[4799]: I1129 05:57:55.883251 4799 generic.go:334] "Generic (PLEG): container finished" podID="6da13658-9a5a-44cd-944d-7bc1891c812c" containerID="f243205ea20722f5af6fc2a958aefd914076ee90a4f0f04c1a0bb2add06aef62" exitCode=0 Nov 29 05:57:55 crc kubenswrapper[4799]: I1129 05:57:55.883507 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5994b" event={"ID":"6da13658-9a5a-44cd-944d-7bc1891c812c","Type":"ContainerDied","Data":"f243205ea20722f5af6fc2a958aefd914076ee90a4f0f04c1a0bb2add06aef62"} Nov 29 05:57:55 crc kubenswrapper[4799]: I1129 05:57:55.886354 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 29 05:57:57 crc kubenswrapper[4799]: I1129 05:57:57.922593 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5994b" event={"ID":"6da13658-9a5a-44cd-944d-7bc1891c812c","Type":"ContainerStarted","Data":"866dca4813c4011fb5dfff2e601fc089e75d591e5f3b3529c1f16f30eb5335eb"} Nov 29 05:57:58 crc kubenswrapper[4799]: I1129 05:57:58.935586 4799 generic.go:334] "Generic (PLEG): container finished" podID="6da13658-9a5a-44cd-944d-7bc1891c812c" containerID="866dca4813c4011fb5dfff2e601fc089e75d591e5f3b3529c1f16f30eb5335eb" exitCode=0 Nov 29 05:57:58 crc kubenswrapper[4799]: I1129 05:57:58.935962 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5994b" event={"ID":"6da13658-9a5a-44cd-944d-7bc1891c812c","Type":"ContainerDied","Data":"866dca4813c4011fb5dfff2e601fc089e75d591e5f3b3529c1f16f30eb5335eb"} Nov 29 05:57:59 crc kubenswrapper[4799]: I1129 05:57:59.946818 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5994b" event={"ID":"6da13658-9a5a-44cd-944d-7bc1891c812c","Type":"ContainerStarted","Data":"5b59acd8ffbbfe451e4f47c8d0ec1e5e2512cab792bc5c36b33e0cbbf2378b04"} Nov 29 05:58:03 crc kubenswrapper[4799]: I1129 05:58:03.379864 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5994b" Nov 29 05:58:03 crc kubenswrapper[4799]: I1129 05:58:03.380175 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5994b" Nov 29 05:58:03 crc kubenswrapper[4799]: I1129 05:58:03.433281 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5994b" Nov 29 05:58:03 crc kubenswrapper[4799]: I1129 05:58:03.449518 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5994b" podStartSLOduration=6.890459553 podStartE2EDuration="10.449498779s" podCreationTimestamp="2025-11-29 05:57:53 +0000 UTC" firstStartedPulling="2025-11-29 05:57:55.886119316 +0000 UTC m=+4751.529049716" lastFinishedPulling="2025-11-29 05:57:59.445158542 +0000 UTC m=+4755.088088942" observedRunningTime="2025-11-29 05:57:59.970616792 +0000 UTC m=+4755.613547192" watchObservedRunningTime="2025-11-29 05:58:03.449498779 +0000 UTC m=+4759.092429179" Nov 29 05:58:07 crc kubenswrapper[4799]: I1129 05:58:07.497729 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 05:58:07 crc kubenswrapper[4799]: I1129 05:58:07.498480 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 05:58:13 crc kubenswrapper[4799]: I1129 05:58:13.426567 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5994b" Nov 29 05:58:13 crc kubenswrapper[4799]: I1129 05:58:13.480910 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5994b"] Nov 29 05:58:14 crc kubenswrapper[4799]: I1129 05:58:14.062548 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5994b" podUID="6da13658-9a5a-44cd-944d-7bc1891c812c" containerName="registry-server" containerID="cri-o://5b59acd8ffbbfe451e4f47c8d0ec1e5e2512cab792bc5c36b33e0cbbf2378b04" gracePeriod=2 Nov 29 05:58:14 crc kubenswrapper[4799]: I1129 05:58:14.509392 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5994b" Nov 29 05:58:14 crc kubenswrapper[4799]: I1129 05:58:14.561236 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6da13658-9a5a-44cd-944d-7bc1891c812c-utilities\") pod \"6da13658-9a5a-44cd-944d-7bc1891c812c\" (UID: \"6da13658-9a5a-44cd-944d-7bc1891c812c\") " Nov 29 05:58:14 crc kubenswrapper[4799]: I1129 05:58:14.561382 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-52bj6\" (UniqueName: \"kubernetes.io/projected/6da13658-9a5a-44cd-944d-7bc1891c812c-kube-api-access-52bj6\") pod \"6da13658-9a5a-44cd-944d-7bc1891c812c\" (UID: \"6da13658-9a5a-44cd-944d-7bc1891c812c\") " Nov 29 05:58:14 crc kubenswrapper[4799]: I1129 05:58:14.561466 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6da13658-9a5a-44cd-944d-7bc1891c812c-catalog-content\") pod \"6da13658-9a5a-44cd-944d-7bc1891c812c\" (UID: \"6da13658-9a5a-44cd-944d-7bc1891c812c\") " Nov 29 05:58:14 crc kubenswrapper[4799]: I1129 05:58:14.562271 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6da13658-9a5a-44cd-944d-7bc1891c812c-utilities" (OuterVolumeSpecName: "utilities") pod "6da13658-9a5a-44cd-944d-7bc1891c812c" (UID: "6da13658-9a5a-44cd-944d-7bc1891c812c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:58:14 crc kubenswrapper[4799]: I1129 05:58:14.573676 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6da13658-9a5a-44cd-944d-7bc1891c812c-kube-api-access-52bj6" (OuterVolumeSpecName: "kube-api-access-52bj6") pod "6da13658-9a5a-44cd-944d-7bc1891c812c" (UID: "6da13658-9a5a-44cd-944d-7bc1891c812c"). InnerVolumeSpecName "kube-api-access-52bj6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:58:14 crc kubenswrapper[4799]: I1129 05:58:14.616106 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6da13658-9a5a-44cd-944d-7bc1891c812c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6da13658-9a5a-44cd-944d-7bc1891c812c" (UID: "6da13658-9a5a-44cd-944d-7bc1891c812c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:58:14 crc kubenswrapper[4799]: I1129 05:58:14.667083 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6da13658-9a5a-44cd-944d-7bc1891c812c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 05:58:14 crc kubenswrapper[4799]: I1129 05:58:14.667118 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6da13658-9a5a-44cd-944d-7bc1891c812c-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 05:58:14 crc kubenswrapper[4799]: I1129 05:58:14.667127 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-52bj6\" (UniqueName: \"kubernetes.io/projected/6da13658-9a5a-44cd-944d-7bc1891c812c-kube-api-access-52bj6\") on node \"crc\" DevicePath \"\"" Nov 29 05:58:15 crc kubenswrapper[4799]: I1129 05:58:15.073317 4799 generic.go:334] "Generic (PLEG): container finished" podID="6da13658-9a5a-44cd-944d-7bc1891c812c" containerID="5b59acd8ffbbfe451e4f47c8d0ec1e5e2512cab792bc5c36b33e0cbbf2378b04" exitCode=0 Nov 29 05:58:15 crc kubenswrapper[4799]: I1129 05:58:15.073622 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5994b" event={"ID":"6da13658-9a5a-44cd-944d-7bc1891c812c","Type":"ContainerDied","Data":"5b59acd8ffbbfe451e4f47c8d0ec1e5e2512cab792bc5c36b33e0cbbf2378b04"} Nov 29 05:58:15 crc kubenswrapper[4799]: I1129 05:58:15.073649 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5994b" event={"ID":"6da13658-9a5a-44cd-944d-7bc1891c812c","Type":"ContainerDied","Data":"88f9923712615b0d6d98d7e60a0fc371f9466e69c7877255fc39d15c406b2502"} Nov 29 05:58:15 crc kubenswrapper[4799]: I1129 05:58:15.073664 4799 scope.go:117] "RemoveContainer" containerID="5b59acd8ffbbfe451e4f47c8d0ec1e5e2512cab792bc5c36b33e0cbbf2378b04" Nov 29 05:58:15 crc kubenswrapper[4799]: I1129 05:58:15.073783 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5994b" Nov 29 05:58:15 crc kubenswrapper[4799]: I1129 05:58:15.100535 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5994b"] Nov 29 05:58:15 crc kubenswrapper[4799]: I1129 05:58:15.114066 4799 scope.go:117] "RemoveContainer" containerID="866dca4813c4011fb5dfff2e601fc089e75d591e5f3b3529c1f16f30eb5335eb" Nov 29 05:58:15 crc kubenswrapper[4799]: I1129 05:58:15.129331 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5994b"] Nov 29 05:58:15 crc kubenswrapper[4799]: I1129 05:58:15.144715 4799 scope.go:117] "RemoveContainer" containerID="f243205ea20722f5af6fc2a958aefd914076ee90a4f0f04c1a0bb2add06aef62" Nov 29 05:58:15 crc kubenswrapper[4799]: I1129 05:58:15.226083 4799 scope.go:117] "RemoveContainer" containerID="5b59acd8ffbbfe451e4f47c8d0ec1e5e2512cab792bc5c36b33e0cbbf2378b04" Nov 29 05:58:15 crc kubenswrapper[4799]: E1129 05:58:15.226566 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b59acd8ffbbfe451e4f47c8d0ec1e5e2512cab792bc5c36b33e0cbbf2378b04\": container with ID starting with 5b59acd8ffbbfe451e4f47c8d0ec1e5e2512cab792bc5c36b33e0cbbf2378b04 not found: ID does not exist" containerID="5b59acd8ffbbfe451e4f47c8d0ec1e5e2512cab792bc5c36b33e0cbbf2378b04" Nov 29 05:58:15 crc kubenswrapper[4799]: I1129 05:58:15.226620 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b59acd8ffbbfe451e4f47c8d0ec1e5e2512cab792bc5c36b33e0cbbf2378b04"} err="failed to get container status \"5b59acd8ffbbfe451e4f47c8d0ec1e5e2512cab792bc5c36b33e0cbbf2378b04\": rpc error: code = NotFound desc = could not find container \"5b59acd8ffbbfe451e4f47c8d0ec1e5e2512cab792bc5c36b33e0cbbf2378b04\": container with ID starting with 5b59acd8ffbbfe451e4f47c8d0ec1e5e2512cab792bc5c36b33e0cbbf2378b04 not found: ID does not exist" Nov 29 05:58:15 crc kubenswrapper[4799]: I1129 05:58:15.226663 4799 scope.go:117] "RemoveContainer" containerID="866dca4813c4011fb5dfff2e601fc089e75d591e5f3b3529c1f16f30eb5335eb" Nov 29 05:58:15 crc kubenswrapper[4799]: E1129 05:58:15.227070 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"866dca4813c4011fb5dfff2e601fc089e75d591e5f3b3529c1f16f30eb5335eb\": container with ID starting with 866dca4813c4011fb5dfff2e601fc089e75d591e5f3b3529c1f16f30eb5335eb not found: ID does not exist" containerID="866dca4813c4011fb5dfff2e601fc089e75d591e5f3b3529c1f16f30eb5335eb" Nov 29 05:58:15 crc kubenswrapper[4799]: I1129 05:58:15.227104 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"866dca4813c4011fb5dfff2e601fc089e75d591e5f3b3529c1f16f30eb5335eb"} err="failed to get container status \"866dca4813c4011fb5dfff2e601fc089e75d591e5f3b3529c1f16f30eb5335eb\": rpc error: code = NotFound desc = could not find container \"866dca4813c4011fb5dfff2e601fc089e75d591e5f3b3529c1f16f30eb5335eb\": container with ID starting with 866dca4813c4011fb5dfff2e601fc089e75d591e5f3b3529c1f16f30eb5335eb not found: ID does not exist" Nov 29 05:58:15 crc kubenswrapper[4799]: I1129 05:58:15.227130 4799 scope.go:117] "RemoveContainer" containerID="f243205ea20722f5af6fc2a958aefd914076ee90a4f0f04c1a0bb2add06aef62" Nov 29 05:58:15 crc kubenswrapper[4799]: E1129 05:58:15.227390 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f243205ea20722f5af6fc2a958aefd914076ee90a4f0f04c1a0bb2add06aef62\": container with ID starting with f243205ea20722f5af6fc2a958aefd914076ee90a4f0f04c1a0bb2add06aef62 not found: ID does not exist" containerID="f243205ea20722f5af6fc2a958aefd914076ee90a4f0f04c1a0bb2add06aef62" Nov 29 05:58:15 crc kubenswrapper[4799]: I1129 05:58:15.227419 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f243205ea20722f5af6fc2a958aefd914076ee90a4f0f04c1a0bb2add06aef62"} err="failed to get container status \"f243205ea20722f5af6fc2a958aefd914076ee90a4f0f04c1a0bb2add06aef62\": rpc error: code = NotFound desc = could not find container \"f243205ea20722f5af6fc2a958aefd914076ee90a4f0f04c1a0bb2add06aef62\": container with ID starting with f243205ea20722f5af6fc2a958aefd914076ee90a4f0f04c1a0bb2add06aef62 not found: ID does not exist" Nov 29 05:58:16 crc kubenswrapper[4799]: I1129 05:58:16.674016 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6da13658-9a5a-44cd-944d-7bc1891c812c" path="/var/lib/kubelet/pods/6da13658-9a5a-44cd-944d-7bc1891c812c/volumes" Nov 29 05:58:37 crc kubenswrapper[4799]: I1129 05:58:37.497540 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 05:58:37 crc kubenswrapper[4799]: I1129 05:58:37.498149 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 05:59:07 crc kubenswrapper[4799]: I1129 05:59:07.498124 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 05:59:07 crc kubenswrapper[4799]: I1129 05:59:07.498657 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 05:59:07 crc kubenswrapper[4799]: I1129 05:59:07.498704 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-26t88" Nov 29 05:59:07 crc kubenswrapper[4799]: I1129 05:59:07.499453 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"53954acbaab2bd75534d88a03d4525340098eb9e1108f075c344f3be17a8ee5e"} pod="openshift-machine-config-operator/machine-config-daemon-26t88" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 29 05:59:07 crc kubenswrapper[4799]: I1129 05:59:07.499539 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" containerID="cri-o://53954acbaab2bd75534d88a03d4525340098eb9e1108f075c344f3be17a8ee5e" gracePeriod=600 Nov 29 05:59:08 crc kubenswrapper[4799]: I1129 05:59:08.596034 4799 generic.go:334] "Generic (PLEG): container finished" podID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerID="53954acbaab2bd75534d88a03d4525340098eb9e1108f075c344f3be17a8ee5e" exitCode=0 Nov 29 05:59:08 crc kubenswrapper[4799]: I1129 05:59:08.596084 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" event={"ID":"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8","Type":"ContainerDied","Data":"53954acbaab2bd75534d88a03d4525340098eb9e1108f075c344f3be17a8ee5e"} Nov 29 05:59:08 crc kubenswrapper[4799]: I1129 05:59:08.596659 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" event={"ID":"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8","Type":"ContainerStarted","Data":"3e8f9ab0a87eb00c33c50fb4c145911e9291e30e11bbe8189bd24ecce45a24f6"} Nov 29 05:59:08 crc kubenswrapper[4799]: I1129 05:59:08.596681 4799 scope.go:117] "RemoveContainer" containerID="b560e06b6de1edd9a7f603575cb604d0893a8a24eda3e04d2446b813f4e889f8" Nov 29 05:59:26 crc kubenswrapper[4799]: I1129 05:59:26.777518 4799 generic.go:334] "Generic (PLEG): container finished" podID="426fd24d-c567-40bf-902b-ce637dbf9f84" containerID="c690a584291d73690cd800625a57f4a66b22b00d44a4b37de8494177dfc7e871" exitCode=0 Nov 29 05:59:26 crc kubenswrapper[4799]: I1129 05:59:26.777944 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t9ms2/must-gather-c5k8l" event={"ID":"426fd24d-c567-40bf-902b-ce637dbf9f84","Type":"ContainerDied","Data":"c690a584291d73690cd800625a57f4a66b22b00d44a4b37de8494177dfc7e871"} Nov 29 05:59:26 crc kubenswrapper[4799]: I1129 05:59:26.778521 4799 scope.go:117] "RemoveContainer" containerID="c690a584291d73690cd800625a57f4a66b22b00d44a4b37de8494177dfc7e871" Nov 29 05:59:27 crc kubenswrapper[4799]: I1129 05:59:27.153817 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-t9ms2_must-gather-c5k8l_426fd24d-c567-40bf-902b-ce637dbf9f84/gather/0.log" Nov 29 05:59:37 crc kubenswrapper[4799]: I1129 05:59:37.402510 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-t9ms2/must-gather-c5k8l"] Nov 29 05:59:37 crc kubenswrapper[4799]: I1129 05:59:37.408081 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-t9ms2/must-gather-c5k8l" podUID="426fd24d-c567-40bf-902b-ce637dbf9f84" containerName="copy" containerID="cri-o://50442bc8a06a2575e37ee856944ca8d1b43b4b863b1de800c9168c4f5fbf45d4" gracePeriod=2 Nov 29 05:59:37 crc kubenswrapper[4799]: I1129 05:59:37.417080 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-t9ms2/must-gather-c5k8l"] Nov 29 05:59:37 crc kubenswrapper[4799]: I1129 05:59:37.832407 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-t9ms2_must-gather-c5k8l_426fd24d-c567-40bf-902b-ce637dbf9f84/copy/0.log" Nov 29 05:59:37 crc kubenswrapper[4799]: I1129 05:59:37.834209 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t9ms2/must-gather-c5k8l" Nov 29 05:59:37 crc kubenswrapper[4799]: I1129 05:59:37.911971 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/426fd24d-c567-40bf-902b-ce637dbf9f84-must-gather-output\") pod \"426fd24d-c567-40bf-902b-ce637dbf9f84\" (UID: \"426fd24d-c567-40bf-902b-ce637dbf9f84\") " Nov 29 05:59:37 crc kubenswrapper[4799]: I1129 05:59:37.912225 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wbgxg\" (UniqueName: \"kubernetes.io/projected/426fd24d-c567-40bf-902b-ce637dbf9f84-kube-api-access-wbgxg\") pod \"426fd24d-c567-40bf-902b-ce637dbf9f84\" (UID: \"426fd24d-c567-40bf-902b-ce637dbf9f84\") " Nov 29 05:59:37 crc kubenswrapper[4799]: I1129 05:59:37.915064 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-t9ms2_must-gather-c5k8l_426fd24d-c567-40bf-902b-ce637dbf9f84/copy/0.log" Nov 29 05:59:37 crc kubenswrapper[4799]: I1129 05:59:37.915620 4799 generic.go:334] "Generic (PLEG): container finished" podID="426fd24d-c567-40bf-902b-ce637dbf9f84" containerID="50442bc8a06a2575e37ee856944ca8d1b43b4b863b1de800c9168c4f5fbf45d4" exitCode=143 Nov 29 05:59:37 crc kubenswrapper[4799]: I1129 05:59:37.915675 4799 scope.go:117] "RemoveContainer" containerID="50442bc8a06a2575e37ee856944ca8d1b43b4b863b1de800c9168c4f5fbf45d4" Nov 29 05:59:37 crc kubenswrapper[4799]: I1129 05:59:37.915824 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t9ms2/must-gather-c5k8l" Nov 29 05:59:37 crc kubenswrapper[4799]: I1129 05:59:37.921230 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/426fd24d-c567-40bf-902b-ce637dbf9f84-kube-api-access-wbgxg" (OuterVolumeSpecName: "kube-api-access-wbgxg") pod "426fd24d-c567-40bf-902b-ce637dbf9f84" (UID: "426fd24d-c567-40bf-902b-ce637dbf9f84"). InnerVolumeSpecName "kube-api-access-wbgxg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 05:59:38 crc kubenswrapper[4799]: I1129 05:59:38.012959 4799 scope.go:117] "RemoveContainer" containerID="c690a584291d73690cd800625a57f4a66b22b00d44a4b37de8494177dfc7e871" Nov 29 05:59:38 crc kubenswrapper[4799]: I1129 05:59:38.018990 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wbgxg\" (UniqueName: \"kubernetes.io/projected/426fd24d-c567-40bf-902b-ce637dbf9f84-kube-api-access-wbgxg\") on node \"crc\" DevicePath \"\"" Nov 29 05:59:38 crc kubenswrapper[4799]: I1129 05:59:38.057624 4799 scope.go:117] "RemoveContainer" containerID="50442bc8a06a2575e37ee856944ca8d1b43b4b863b1de800c9168c4f5fbf45d4" Nov 29 05:59:38 crc kubenswrapper[4799]: E1129 05:59:38.057997 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50442bc8a06a2575e37ee856944ca8d1b43b4b863b1de800c9168c4f5fbf45d4\": container with ID starting with 50442bc8a06a2575e37ee856944ca8d1b43b4b863b1de800c9168c4f5fbf45d4 not found: ID does not exist" containerID="50442bc8a06a2575e37ee856944ca8d1b43b4b863b1de800c9168c4f5fbf45d4" Nov 29 05:59:38 crc kubenswrapper[4799]: I1129 05:59:38.058105 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50442bc8a06a2575e37ee856944ca8d1b43b4b863b1de800c9168c4f5fbf45d4"} err="failed to get container status \"50442bc8a06a2575e37ee856944ca8d1b43b4b863b1de800c9168c4f5fbf45d4\": rpc error: code = NotFound desc = could not find container \"50442bc8a06a2575e37ee856944ca8d1b43b4b863b1de800c9168c4f5fbf45d4\": container with ID starting with 50442bc8a06a2575e37ee856944ca8d1b43b4b863b1de800c9168c4f5fbf45d4 not found: ID does not exist" Nov 29 05:59:38 crc kubenswrapper[4799]: I1129 05:59:38.058132 4799 scope.go:117] "RemoveContainer" containerID="c690a584291d73690cd800625a57f4a66b22b00d44a4b37de8494177dfc7e871" Nov 29 05:59:38 crc kubenswrapper[4799]: E1129 05:59:38.059612 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c690a584291d73690cd800625a57f4a66b22b00d44a4b37de8494177dfc7e871\": container with ID starting with c690a584291d73690cd800625a57f4a66b22b00d44a4b37de8494177dfc7e871 not found: ID does not exist" containerID="c690a584291d73690cd800625a57f4a66b22b00d44a4b37de8494177dfc7e871" Nov 29 05:59:38 crc kubenswrapper[4799]: I1129 05:59:38.059638 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c690a584291d73690cd800625a57f4a66b22b00d44a4b37de8494177dfc7e871"} err="failed to get container status \"c690a584291d73690cd800625a57f4a66b22b00d44a4b37de8494177dfc7e871\": rpc error: code = NotFound desc = could not find container \"c690a584291d73690cd800625a57f4a66b22b00d44a4b37de8494177dfc7e871\": container with ID starting with c690a584291d73690cd800625a57f4a66b22b00d44a4b37de8494177dfc7e871 not found: ID does not exist" Nov 29 05:59:38 crc kubenswrapper[4799]: I1129 05:59:38.090620 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/426fd24d-c567-40bf-902b-ce637dbf9f84-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "426fd24d-c567-40bf-902b-ce637dbf9f84" (UID: "426fd24d-c567-40bf-902b-ce637dbf9f84"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 05:59:38 crc kubenswrapper[4799]: I1129 05:59:38.122243 4799 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/426fd24d-c567-40bf-902b-ce637dbf9f84-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 29 05:59:38 crc kubenswrapper[4799]: I1129 05:59:38.669586 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="426fd24d-c567-40bf-902b-ce637dbf9f84" path="/var/lib/kubelet/pods/426fd24d-c567-40bf-902b-ce637dbf9f84/volumes" Nov 29 05:59:59 crc kubenswrapper[4799]: I1129 05:59:59.620346 4799 scope.go:117] "RemoveContainer" containerID="499d0333bd30be0734900a58a7daf348badb369ce4c1cb2598b375bd380a1b12" Nov 29 06:00:00 crc kubenswrapper[4799]: I1129 06:00:00.152927 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406600-ghh7p"] Nov 29 06:00:00 crc kubenswrapper[4799]: E1129 06:00:00.153714 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6da13658-9a5a-44cd-944d-7bc1891c812c" containerName="registry-server" Nov 29 06:00:00 crc kubenswrapper[4799]: I1129 06:00:00.153743 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="6da13658-9a5a-44cd-944d-7bc1891c812c" containerName="registry-server" Nov 29 06:00:00 crc kubenswrapper[4799]: E1129 06:00:00.153761 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="426fd24d-c567-40bf-902b-ce637dbf9f84" containerName="gather" Nov 29 06:00:00 crc kubenswrapper[4799]: I1129 06:00:00.153770 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="426fd24d-c567-40bf-902b-ce637dbf9f84" containerName="gather" Nov 29 06:00:00 crc kubenswrapper[4799]: E1129 06:00:00.153915 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6da13658-9a5a-44cd-944d-7bc1891c812c" containerName="extract-utilities" Nov 29 06:00:00 crc kubenswrapper[4799]: I1129 06:00:00.153928 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="6da13658-9a5a-44cd-944d-7bc1891c812c" containerName="extract-utilities" Nov 29 06:00:00 crc kubenswrapper[4799]: E1129 06:00:00.153956 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="426fd24d-c567-40bf-902b-ce637dbf9f84" containerName="copy" Nov 29 06:00:00 crc kubenswrapper[4799]: I1129 06:00:00.153965 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="426fd24d-c567-40bf-902b-ce637dbf9f84" containerName="copy" Nov 29 06:00:00 crc kubenswrapper[4799]: E1129 06:00:00.153982 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6da13658-9a5a-44cd-944d-7bc1891c812c" containerName="extract-content" Nov 29 06:00:00 crc kubenswrapper[4799]: I1129 06:00:00.153990 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="6da13658-9a5a-44cd-944d-7bc1891c812c" containerName="extract-content" Nov 29 06:00:00 crc kubenswrapper[4799]: I1129 06:00:00.154209 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="426fd24d-c567-40bf-902b-ce637dbf9f84" containerName="copy" Nov 29 06:00:00 crc kubenswrapper[4799]: I1129 06:00:00.154236 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="6da13658-9a5a-44cd-944d-7bc1891c812c" containerName="registry-server" Nov 29 06:00:00 crc kubenswrapper[4799]: I1129 06:00:00.154250 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="426fd24d-c567-40bf-902b-ce637dbf9f84" containerName="gather" Nov 29 06:00:00 crc kubenswrapper[4799]: I1129 06:00:00.155095 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406600-ghh7p" Nov 29 06:00:00 crc kubenswrapper[4799]: I1129 06:00:00.157491 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 29 06:00:00 crc kubenswrapper[4799]: I1129 06:00:00.159516 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 29 06:00:00 crc kubenswrapper[4799]: I1129 06:00:00.165244 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406600-ghh7p"] Nov 29 06:00:00 crc kubenswrapper[4799]: I1129 06:00:00.292330 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtngc\" (UniqueName: \"kubernetes.io/projected/85b1cf0d-deee-4f81-87b4-ba9f269b9acc-kube-api-access-xtngc\") pod \"collect-profiles-29406600-ghh7p\" (UID: \"85b1cf0d-deee-4f81-87b4-ba9f269b9acc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406600-ghh7p" Nov 29 06:00:00 crc kubenswrapper[4799]: I1129 06:00:00.292422 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/85b1cf0d-deee-4f81-87b4-ba9f269b9acc-secret-volume\") pod \"collect-profiles-29406600-ghh7p\" (UID: \"85b1cf0d-deee-4f81-87b4-ba9f269b9acc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406600-ghh7p" Nov 29 06:00:00 crc kubenswrapper[4799]: I1129 06:00:00.292482 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/85b1cf0d-deee-4f81-87b4-ba9f269b9acc-config-volume\") pod \"collect-profiles-29406600-ghh7p\" (UID: \"85b1cf0d-deee-4f81-87b4-ba9f269b9acc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406600-ghh7p" Nov 29 06:00:00 crc kubenswrapper[4799]: I1129 06:00:00.396040 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtngc\" (UniqueName: \"kubernetes.io/projected/85b1cf0d-deee-4f81-87b4-ba9f269b9acc-kube-api-access-xtngc\") pod \"collect-profiles-29406600-ghh7p\" (UID: \"85b1cf0d-deee-4f81-87b4-ba9f269b9acc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406600-ghh7p" Nov 29 06:00:00 crc kubenswrapper[4799]: I1129 06:00:00.396123 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/85b1cf0d-deee-4f81-87b4-ba9f269b9acc-secret-volume\") pod \"collect-profiles-29406600-ghh7p\" (UID: \"85b1cf0d-deee-4f81-87b4-ba9f269b9acc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406600-ghh7p" Nov 29 06:00:00 crc kubenswrapper[4799]: I1129 06:00:00.396298 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/85b1cf0d-deee-4f81-87b4-ba9f269b9acc-config-volume\") pod \"collect-profiles-29406600-ghh7p\" (UID: \"85b1cf0d-deee-4f81-87b4-ba9f269b9acc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406600-ghh7p" Nov 29 06:00:00 crc kubenswrapper[4799]: I1129 06:00:00.397180 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/85b1cf0d-deee-4f81-87b4-ba9f269b9acc-config-volume\") pod \"collect-profiles-29406600-ghh7p\" (UID: \"85b1cf0d-deee-4f81-87b4-ba9f269b9acc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406600-ghh7p" Nov 29 06:00:00 crc kubenswrapper[4799]: I1129 06:00:00.408723 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/85b1cf0d-deee-4f81-87b4-ba9f269b9acc-secret-volume\") pod \"collect-profiles-29406600-ghh7p\" (UID: \"85b1cf0d-deee-4f81-87b4-ba9f269b9acc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406600-ghh7p" Nov 29 06:00:00 crc kubenswrapper[4799]: I1129 06:00:00.418327 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtngc\" (UniqueName: \"kubernetes.io/projected/85b1cf0d-deee-4f81-87b4-ba9f269b9acc-kube-api-access-xtngc\") pod \"collect-profiles-29406600-ghh7p\" (UID: \"85b1cf0d-deee-4f81-87b4-ba9f269b9acc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406600-ghh7p" Nov 29 06:00:00 crc kubenswrapper[4799]: I1129 06:00:00.475318 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406600-ghh7p" Nov 29 06:00:00 crc kubenswrapper[4799]: I1129 06:00:00.914218 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406600-ghh7p"] Nov 29 06:00:01 crc kubenswrapper[4799]: I1129 06:00:01.162448 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29406600-ghh7p" event={"ID":"85b1cf0d-deee-4f81-87b4-ba9f269b9acc","Type":"ContainerStarted","Data":"8eca89699093b641d2efeb89b0fc0b0f9a8447289c09694cea8a5353de68e72f"} Nov 29 06:00:01 crc kubenswrapper[4799]: I1129 06:00:01.162740 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29406600-ghh7p" event={"ID":"85b1cf0d-deee-4f81-87b4-ba9f269b9acc","Type":"ContainerStarted","Data":"2a5dbe148eccff81d349b479733d655a5988d2a5f11322df02c798557844afa4"} Nov 29 06:00:02 crc kubenswrapper[4799]: I1129 06:00:02.185514 4799 generic.go:334] "Generic (PLEG): container finished" podID="85b1cf0d-deee-4f81-87b4-ba9f269b9acc" containerID="8eca89699093b641d2efeb89b0fc0b0f9a8447289c09694cea8a5353de68e72f" exitCode=0 Nov 29 06:00:02 crc kubenswrapper[4799]: I1129 06:00:02.186208 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29406600-ghh7p" event={"ID":"85b1cf0d-deee-4f81-87b4-ba9f269b9acc","Type":"ContainerDied","Data":"8eca89699093b641d2efeb89b0fc0b0f9a8447289c09694cea8a5353de68e72f"} Nov 29 06:00:03 crc kubenswrapper[4799]: I1129 06:00:03.484077 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406600-ghh7p" Nov 29 06:00:03 crc kubenswrapper[4799]: I1129 06:00:03.656839 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xtngc\" (UniqueName: \"kubernetes.io/projected/85b1cf0d-deee-4f81-87b4-ba9f269b9acc-kube-api-access-xtngc\") pod \"85b1cf0d-deee-4f81-87b4-ba9f269b9acc\" (UID: \"85b1cf0d-deee-4f81-87b4-ba9f269b9acc\") " Nov 29 06:00:03 crc kubenswrapper[4799]: I1129 06:00:03.656902 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/85b1cf0d-deee-4f81-87b4-ba9f269b9acc-config-volume\") pod \"85b1cf0d-deee-4f81-87b4-ba9f269b9acc\" (UID: \"85b1cf0d-deee-4f81-87b4-ba9f269b9acc\") " Nov 29 06:00:03 crc kubenswrapper[4799]: I1129 06:00:03.656992 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/85b1cf0d-deee-4f81-87b4-ba9f269b9acc-secret-volume\") pod \"85b1cf0d-deee-4f81-87b4-ba9f269b9acc\" (UID: \"85b1cf0d-deee-4f81-87b4-ba9f269b9acc\") " Nov 29 06:00:03 crc kubenswrapper[4799]: I1129 06:00:03.657976 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85b1cf0d-deee-4f81-87b4-ba9f269b9acc-config-volume" (OuterVolumeSpecName: "config-volume") pod "85b1cf0d-deee-4f81-87b4-ba9f269b9acc" (UID: "85b1cf0d-deee-4f81-87b4-ba9f269b9acc"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 06:00:03 crc kubenswrapper[4799]: I1129 06:00:03.762464 4799 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/85b1cf0d-deee-4f81-87b4-ba9f269b9acc-config-volume\") on node \"crc\" DevicePath \"\"" Nov 29 06:00:04 crc kubenswrapper[4799]: I1129 06:00:04.012354 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85b1cf0d-deee-4f81-87b4-ba9f269b9acc-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "85b1cf0d-deee-4f81-87b4-ba9f269b9acc" (UID: "85b1cf0d-deee-4f81-87b4-ba9f269b9acc"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 06:00:04 crc kubenswrapper[4799]: I1129 06:00:04.012695 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85b1cf0d-deee-4f81-87b4-ba9f269b9acc-kube-api-access-xtngc" (OuterVolumeSpecName: "kube-api-access-xtngc") pod "85b1cf0d-deee-4f81-87b4-ba9f269b9acc" (UID: "85b1cf0d-deee-4f81-87b4-ba9f269b9acc"). InnerVolumeSpecName "kube-api-access-xtngc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 06:00:04 crc kubenswrapper[4799]: I1129 06:00:04.069718 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xtngc\" (UniqueName: \"kubernetes.io/projected/85b1cf0d-deee-4f81-87b4-ba9f269b9acc-kube-api-access-xtngc\") on node \"crc\" DevicePath \"\"" Nov 29 06:00:04 crc kubenswrapper[4799]: I1129 06:00:04.069764 4799 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/85b1cf0d-deee-4f81-87b4-ba9f269b9acc-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 29 06:00:04 crc kubenswrapper[4799]: I1129 06:00:04.203814 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29406600-ghh7p" event={"ID":"85b1cf0d-deee-4f81-87b4-ba9f269b9acc","Type":"ContainerDied","Data":"2a5dbe148eccff81d349b479733d655a5988d2a5f11322df02c798557844afa4"} Nov 29 06:00:04 crc kubenswrapper[4799]: I1129 06:00:04.204107 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a5dbe148eccff81d349b479733d655a5988d2a5f11322df02c798557844afa4" Nov 29 06:00:04 crc kubenswrapper[4799]: I1129 06:00:04.204155 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406600-ghh7p" Nov 29 06:00:04 crc kubenswrapper[4799]: I1129 06:00:04.249512 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406555-g8t85"] Nov 29 06:00:04 crc kubenswrapper[4799]: I1129 06:00:04.256605 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406555-g8t85"] Nov 29 06:00:04 crc kubenswrapper[4799]: I1129 06:00:04.673908 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2dad815-9bd5-4c9f-957a-16ed0df1ff5e" path="/var/lib/kubelet/pods/a2dad815-9bd5-4c9f-957a-16ed0df1ff5e/volumes" Nov 29 06:00:59 crc kubenswrapper[4799]: I1129 06:00:59.683638 4799 scope.go:117] "RemoveContainer" containerID="cd5f1d2858843448767b327c15763b00e79599889269e9b82425990c734cba43" Nov 29 06:00:59 crc kubenswrapper[4799]: I1129 06:00:59.706804 4799 scope.go:117] "RemoveContainer" containerID="fe872e5362e75db2c42487603d95f07dc9590830840138f3bf0c3f0d4085ae2c" Nov 29 06:01:00 crc kubenswrapper[4799]: I1129 06:01:00.150224 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29406601-vwx4k"] Nov 29 06:01:00 crc kubenswrapper[4799]: E1129 06:01:00.151053 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85b1cf0d-deee-4f81-87b4-ba9f269b9acc" containerName="collect-profiles" Nov 29 06:01:00 crc kubenswrapper[4799]: I1129 06:01:00.151074 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="85b1cf0d-deee-4f81-87b4-ba9f269b9acc" containerName="collect-profiles" Nov 29 06:01:00 crc kubenswrapper[4799]: I1129 06:01:00.151324 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="85b1cf0d-deee-4f81-87b4-ba9f269b9acc" containerName="collect-profiles" Nov 29 06:01:00 crc kubenswrapper[4799]: I1129 06:01:00.152081 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29406601-vwx4k" Nov 29 06:01:00 crc kubenswrapper[4799]: I1129 06:01:00.176864 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29406601-vwx4k"] Nov 29 06:01:00 crc kubenswrapper[4799]: I1129 06:01:00.347014 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cddc74cf-2d7e-4099-90bb-fc34b38063b5-combined-ca-bundle\") pod \"keystone-cron-29406601-vwx4k\" (UID: \"cddc74cf-2d7e-4099-90bb-fc34b38063b5\") " pod="openstack/keystone-cron-29406601-vwx4k" Nov 29 06:01:00 crc kubenswrapper[4799]: I1129 06:01:00.347066 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cddc74cf-2d7e-4099-90bb-fc34b38063b5-config-data\") pod \"keystone-cron-29406601-vwx4k\" (UID: \"cddc74cf-2d7e-4099-90bb-fc34b38063b5\") " pod="openstack/keystone-cron-29406601-vwx4k" Nov 29 06:01:00 crc kubenswrapper[4799]: I1129 06:01:00.347190 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwtl7\" (UniqueName: \"kubernetes.io/projected/cddc74cf-2d7e-4099-90bb-fc34b38063b5-kube-api-access-pwtl7\") pod \"keystone-cron-29406601-vwx4k\" (UID: \"cddc74cf-2d7e-4099-90bb-fc34b38063b5\") " pod="openstack/keystone-cron-29406601-vwx4k" Nov 29 06:01:00 crc kubenswrapper[4799]: I1129 06:01:00.347220 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cddc74cf-2d7e-4099-90bb-fc34b38063b5-fernet-keys\") pod \"keystone-cron-29406601-vwx4k\" (UID: \"cddc74cf-2d7e-4099-90bb-fc34b38063b5\") " pod="openstack/keystone-cron-29406601-vwx4k" Nov 29 06:01:00 crc kubenswrapper[4799]: I1129 06:01:00.449113 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwtl7\" (UniqueName: \"kubernetes.io/projected/cddc74cf-2d7e-4099-90bb-fc34b38063b5-kube-api-access-pwtl7\") pod \"keystone-cron-29406601-vwx4k\" (UID: \"cddc74cf-2d7e-4099-90bb-fc34b38063b5\") " pod="openstack/keystone-cron-29406601-vwx4k" Nov 29 06:01:00 crc kubenswrapper[4799]: I1129 06:01:00.449175 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cddc74cf-2d7e-4099-90bb-fc34b38063b5-fernet-keys\") pod \"keystone-cron-29406601-vwx4k\" (UID: \"cddc74cf-2d7e-4099-90bb-fc34b38063b5\") " pod="openstack/keystone-cron-29406601-vwx4k" Nov 29 06:01:00 crc kubenswrapper[4799]: I1129 06:01:00.449225 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cddc74cf-2d7e-4099-90bb-fc34b38063b5-combined-ca-bundle\") pod \"keystone-cron-29406601-vwx4k\" (UID: \"cddc74cf-2d7e-4099-90bb-fc34b38063b5\") " pod="openstack/keystone-cron-29406601-vwx4k" Nov 29 06:01:00 crc kubenswrapper[4799]: I1129 06:01:00.449247 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cddc74cf-2d7e-4099-90bb-fc34b38063b5-config-data\") pod \"keystone-cron-29406601-vwx4k\" (UID: \"cddc74cf-2d7e-4099-90bb-fc34b38063b5\") " pod="openstack/keystone-cron-29406601-vwx4k" Nov 29 06:01:00 crc kubenswrapper[4799]: I1129 06:01:00.456319 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cddc74cf-2d7e-4099-90bb-fc34b38063b5-fernet-keys\") pod \"keystone-cron-29406601-vwx4k\" (UID: \"cddc74cf-2d7e-4099-90bb-fc34b38063b5\") " pod="openstack/keystone-cron-29406601-vwx4k" Nov 29 06:01:00 crc kubenswrapper[4799]: I1129 06:01:00.456426 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cddc74cf-2d7e-4099-90bb-fc34b38063b5-config-data\") pod \"keystone-cron-29406601-vwx4k\" (UID: \"cddc74cf-2d7e-4099-90bb-fc34b38063b5\") " pod="openstack/keystone-cron-29406601-vwx4k" Nov 29 06:01:00 crc kubenswrapper[4799]: I1129 06:01:00.466879 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cddc74cf-2d7e-4099-90bb-fc34b38063b5-combined-ca-bundle\") pod \"keystone-cron-29406601-vwx4k\" (UID: \"cddc74cf-2d7e-4099-90bb-fc34b38063b5\") " pod="openstack/keystone-cron-29406601-vwx4k" Nov 29 06:01:00 crc kubenswrapper[4799]: I1129 06:01:00.477780 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwtl7\" (UniqueName: \"kubernetes.io/projected/cddc74cf-2d7e-4099-90bb-fc34b38063b5-kube-api-access-pwtl7\") pod \"keystone-cron-29406601-vwx4k\" (UID: \"cddc74cf-2d7e-4099-90bb-fc34b38063b5\") " pod="openstack/keystone-cron-29406601-vwx4k" Nov 29 06:01:00 crc kubenswrapper[4799]: I1129 06:01:00.770733 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29406601-vwx4k" Nov 29 06:01:02 crc kubenswrapper[4799]: I1129 06:01:02.065780 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29406601-vwx4k"] Nov 29 06:01:02 crc kubenswrapper[4799]: I1129 06:01:02.225053 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29406601-vwx4k" event={"ID":"cddc74cf-2d7e-4099-90bb-fc34b38063b5","Type":"ContainerStarted","Data":"2a665f0d4ca91856bc84c379e81270bcf6bcd3a4cc1337ce1bdaabb366a58875"} Nov 29 06:01:03 crc kubenswrapper[4799]: I1129 06:01:03.236194 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29406601-vwx4k" event={"ID":"cddc74cf-2d7e-4099-90bb-fc34b38063b5","Type":"ContainerStarted","Data":"f1bae95aba7cec0bac5ee0ff9ea38c7a2b4dfc06765ab00150198126e654bad9"} Nov 29 06:01:03 crc kubenswrapper[4799]: I1129 06:01:03.262371 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29406601-vwx4k" podStartSLOduration=3.262338447 podStartE2EDuration="3.262338447s" podCreationTimestamp="2025-11-29 06:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 06:01:03.251546615 +0000 UTC m=+4938.894477005" watchObservedRunningTime="2025-11-29 06:01:03.262338447 +0000 UTC m=+4938.905268897" Nov 29 06:01:05 crc kubenswrapper[4799]: I1129 06:01:05.256055 4799 generic.go:334] "Generic (PLEG): container finished" podID="cddc74cf-2d7e-4099-90bb-fc34b38063b5" containerID="f1bae95aba7cec0bac5ee0ff9ea38c7a2b4dfc06765ab00150198126e654bad9" exitCode=0 Nov 29 06:01:05 crc kubenswrapper[4799]: I1129 06:01:05.256516 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29406601-vwx4k" event={"ID":"cddc74cf-2d7e-4099-90bb-fc34b38063b5","Type":"ContainerDied","Data":"f1bae95aba7cec0bac5ee0ff9ea38c7a2b4dfc06765ab00150198126e654bad9"} Nov 29 06:01:06 crc kubenswrapper[4799]: I1129 06:01:06.614081 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29406601-vwx4k" Nov 29 06:01:06 crc kubenswrapper[4799]: I1129 06:01:06.790336 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cddc74cf-2d7e-4099-90bb-fc34b38063b5-config-data\") pod \"cddc74cf-2d7e-4099-90bb-fc34b38063b5\" (UID: \"cddc74cf-2d7e-4099-90bb-fc34b38063b5\") " Nov 29 06:01:06 crc kubenswrapper[4799]: I1129 06:01:06.790391 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cddc74cf-2d7e-4099-90bb-fc34b38063b5-combined-ca-bundle\") pod \"cddc74cf-2d7e-4099-90bb-fc34b38063b5\" (UID: \"cddc74cf-2d7e-4099-90bb-fc34b38063b5\") " Nov 29 06:01:06 crc kubenswrapper[4799]: I1129 06:01:06.790468 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pwtl7\" (UniqueName: \"kubernetes.io/projected/cddc74cf-2d7e-4099-90bb-fc34b38063b5-kube-api-access-pwtl7\") pod \"cddc74cf-2d7e-4099-90bb-fc34b38063b5\" (UID: \"cddc74cf-2d7e-4099-90bb-fc34b38063b5\") " Nov 29 06:01:06 crc kubenswrapper[4799]: I1129 06:01:06.790571 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cddc74cf-2d7e-4099-90bb-fc34b38063b5-fernet-keys\") pod \"cddc74cf-2d7e-4099-90bb-fc34b38063b5\" (UID: \"cddc74cf-2d7e-4099-90bb-fc34b38063b5\") " Nov 29 06:01:06 crc kubenswrapper[4799]: I1129 06:01:06.797279 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cddc74cf-2d7e-4099-90bb-fc34b38063b5-kube-api-access-pwtl7" (OuterVolumeSpecName: "kube-api-access-pwtl7") pod "cddc74cf-2d7e-4099-90bb-fc34b38063b5" (UID: "cddc74cf-2d7e-4099-90bb-fc34b38063b5"). InnerVolumeSpecName "kube-api-access-pwtl7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 06:01:06 crc kubenswrapper[4799]: I1129 06:01:06.797308 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cddc74cf-2d7e-4099-90bb-fc34b38063b5-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "cddc74cf-2d7e-4099-90bb-fc34b38063b5" (UID: "cddc74cf-2d7e-4099-90bb-fc34b38063b5"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 06:01:06 crc kubenswrapper[4799]: I1129 06:01:06.836857 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cddc74cf-2d7e-4099-90bb-fc34b38063b5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cddc74cf-2d7e-4099-90bb-fc34b38063b5" (UID: "cddc74cf-2d7e-4099-90bb-fc34b38063b5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 06:01:06 crc kubenswrapper[4799]: I1129 06:01:06.854893 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cddc74cf-2d7e-4099-90bb-fc34b38063b5-config-data" (OuterVolumeSpecName: "config-data") pod "cddc74cf-2d7e-4099-90bb-fc34b38063b5" (UID: "cddc74cf-2d7e-4099-90bb-fc34b38063b5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 06:01:06 crc kubenswrapper[4799]: I1129 06:01:06.903205 4799 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cddc74cf-2d7e-4099-90bb-fc34b38063b5-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 29 06:01:06 crc kubenswrapper[4799]: I1129 06:01:06.903235 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cddc74cf-2d7e-4099-90bb-fc34b38063b5-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 06:01:06 crc kubenswrapper[4799]: I1129 06:01:06.903243 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cddc74cf-2d7e-4099-90bb-fc34b38063b5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 06:01:06 crc kubenswrapper[4799]: I1129 06:01:06.903254 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pwtl7\" (UniqueName: \"kubernetes.io/projected/cddc74cf-2d7e-4099-90bb-fc34b38063b5-kube-api-access-pwtl7\") on node \"crc\" DevicePath \"\"" Nov 29 06:01:07 crc kubenswrapper[4799]: I1129 06:01:07.282000 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29406601-vwx4k" event={"ID":"cddc74cf-2d7e-4099-90bb-fc34b38063b5","Type":"ContainerDied","Data":"2a665f0d4ca91856bc84c379e81270bcf6bcd3a4cc1337ce1bdaabb366a58875"} Nov 29 06:01:07 crc kubenswrapper[4799]: I1129 06:01:07.282091 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a665f0d4ca91856bc84c379e81270bcf6bcd3a4cc1337ce1bdaabb366a58875" Nov 29 06:01:07 crc kubenswrapper[4799]: I1129 06:01:07.282092 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29406601-vwx4k" Nov 29 06:01:07 crc kubenswrapper[4799]: I1129 06:01:07.497208 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 06:01:07 crc kubenswrapper[4799]: I1129 06:01:07.497274 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 06:01:37 crc kubenswrapper[4799]: I1129 06:01:37.498066 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 06:01:37 crc kubenswrapper[4799]: I1129 06:01:37.498658 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 06:02:07 crc kubenswrapper[4799]: I1129 06:02:07.497382 4799 patch_prober.go:28] interesting pod/machine-config-daemon-26t88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 06:02:07 crc kubenswrapper[4799]: I1129 06:02:07.498146 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 06:02:07 crc kubenswrapper[4799]: I1129 06:02:07.498212 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-26t88" Nov 29 06:02:07 crc kubenswrapper[4799]: I1129 06:02:07.499142 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3e8f9ab0a87eb00c33c50fb4c145911e9291e30e11bbe8189bd24ecce45a24f6"} pod="openshift-machine-config-operator/machine-config-daemon-26t88" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 29 06:02:07 crc kubenswrapper[4799]: I1129 06:02:07.499225 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerName="machine-config-daemon" containerID="cri-o://3e8f9ab0a87eb00c33c50fb4c145911e9291e30e11bbe8189bd24ecce45a24f6" gracePeriod=600 Nov 29 06:02:08 crc kubenswrapper[4799]: E1129 06:02:08.642162 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 06:02:08 crc kubenswrapper[4799]: I1129 06:02:08.884906 4799 generic.go:334] "Generic (PLEG): container finished" podID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" containerID="3e8f9ab0a87eb00c33c50fb4c145911e9291e30e11bbe8189bd24ecce45a24f6" exitCode=0 Nov 29 06:02:08 crc kubenswrapper[4799]: I1129 06:02:08.885274 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-26t88" event={"ID":"b333a2e0-0da7-4d88-9539-0b4cc12bc2e8","Type":"ContainerDied","Data":"3e8f9ab0a87eb00c33c50fb4c145911e9291e30e11bbe8189bd24ecce45a24f6"} Nov 29 06:02:08 crc kubenswrapper[4799]: I1129 06:02:08.885354 4799 scope.go:117] "RemoveContainer" containerID="53954acbaab2bd75534d88a03d4525340098eb9e1108f075c344f3be17a8ee5e" Nov 29 06:02:08 crc kubenswrapper[4799]: I1129 06:02:08.886143 4799 scope.go:117] "RemoveContainer" containerID="3e8f9ab0a87eb00c33c50fb4c145911e9291e30e11bbe8189bd24ecce45a24f6" Nov 29 06:02:08 crc kubenswrapper[4799]: E1129 06:02:08.886447 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 06:02:23 crc kubenswrapper[4799]: I1129 06:02:23.660094 4799 scope.go:117] "RemoveContainer" containerID="3e8f9ab0a87eb00c33c50fb4c145911e9291e30e11bbe8189bd24ecce45a24f6" Nov 29 06:02:23 crc kubenswrapper[4799]: E1129 06:02:23.661030 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 06:02:34 crc kubenswrapper[4799]: I1129 06:02:34.666091 4799 scope.go:117] "RemoveContainer" containerID="3e8f9ab0a87eb00c33c50fb4c145911e9291e30e11bbe8189bd24ecce45a24f6" Nov 29 06:02:34 crc kubenswrapper[4799]: E1129 06:02:34.667188 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 06:02:48 crc kubenswrapper[4799]: I1129 06:02:48.659274 4799 scope.go:117] "RemoveContainer" containerID="3e8f9ab0a87eb00c33c50fb4c145911e9291e30e11bbe8189bd24ecce45a24f6" Nov 29 06:02:48 crc kubenswrapper[4799]: E1129 06:02:48.661482 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" Nov 29 06:03:03 crc kubenswrapper[4799]: I1129 06:03:03.659526 4799 scope.go:117] "RemoveContainer" containerID="3e8f9ab0a87eb00c33c50fb4c145911e9291e30e11bbe8189bd24ecce45a24f6" Nov 29 06:03:03 crc kubenswrapper[4799]: E1129 06:03:03.660439 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-26t88_openshift-machine-config-operator(b333a2e0-0da7-4d88-9539-0b4cc12bc2e8)\"" pod="openshift-machine-config-operator/machine-config-daemon-26t88" podUID="b333a2e0-0da7-4d88-9539-0b4cc12bc2e8" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515112506245024446 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015112506245017363 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015112474022016503 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015112474022015453 5ustar corecore